LawBench

Legal model evaluator

Evaluates the legal knowledge of large language models using a custom benchmarking framework.

Benchmarking Legal Knowledge of Large Language Models

GitHub

273 stars
7 watching
44 forks
Language: Python
last commit: about 1 year ago
benchmarkchatgptlawllm

Related projects:

Repository Description Stars
open-compass/vlmevalkit An evaluation toolkit for large vision-language models 1,514
freedomintelligence/mllm-bench Evaluates and compares the performance of multimodal large language models on various tasks 56
mlabonne/llm-autoeval A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. 566
qcri/llmebench A benchmarking framework for large language models 81
liuhc0428/law-gpt A Chinese law-focused conversational AI model designed to provide reliable and professional legal answers. 1,072
andrewzhe/lawyer-llama An AI model trained on legal data to provide answers and explanations in Chinese law 871
obss/jury A comprehensive toolkit for evaluating NLP experiments offering automated metrics and efficient computation. 187
iclrandd/blackstone Develops an NLP pipeline and model for processing long-form legal text 641
open-compass/mmbench A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models. 168
oeg-upm/lubm4obda Evaluates Ontology-Based Data Access systems with inference and meta knowledge benchmarking 4
openai/simple-evals Evaluates language models using standardized benchmarks and prompting techniques. 2,059
siat-nlp/hanfei Develops and trains a large-scale, parameterized model for legal question answering and text generation 105
maluuba/nlg-eval A toolset for evaluating and comparing natural language generation models 1,350
openlmlab/gaokao-bench An evaluation framework using Chinese high school examination questions to assess large language model capabilities 565
mlgroupjlu/llm-eval-survey A repository of papers and resources for evaluating large language models. 1,450