LawBench
Legal model evaluator
Evaluates the legal knowledge of large language models using a custom benchmarking framework.
Benchmarking Legal Knowledge of Large Language Models
273 stars
7 watching
44 forks
Language: Python
last commit: about 1 year ago benchmarkchatgptlawllm
Related projects:
Repository | Description | Stars |
---|---|---|
open-compass/vlmevalkit | An evaluation toolkit for large vision-language models | 1,514 |
freedomintelligence/mllm-bench | Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
mlabonne/llm-autoeval | A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. | 566 |
qcri/llmebench | A benchmarking framework for large language models | 81 |
liuhc0428/law-gpt | A Chinese law-focused conversational AI model designed to provide reliable and professional legal answers. | 1,072 |
andrewzhe/lawyer-llama | An AI model trained on legal data to provide answers and explanations in Chinese law | 871 |
obss/jury | A comprehensive toolkit for evaluating NLP experiments offering automated metrics and efficient computation. | 187 |
iclrandd/blackstone | Develops an NLP pipeline and model for processing long-form legal text | 641 |
open-compass/mmbench | A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models. | 168 |
oeg-upm/lubm4obda | Evaluates Ontology-Based Data Access systems with inference and meta knowledge benchmarking | 4 |
openai/simple-evals | Evaluates language models using standardized benchmarks and prompting techniques. | 2,059 |
siat-nlp/hanfei | Develops and trains a large-scale, parameterized model for legal question answering and text generation | 105 |
maluuba/nlg-eval | A toolset for evaluating and comparing natural language generation models | 1,350 |
openlmlab/gaokao-bench | An evaluation framework using Chinese high school examination questions to assess large language model capabilities | 565 |
mlgroupjlu/llm-eval-survey | A repository of papers and resources for evaluating large language models. | 1,450 |