LawBench
Legal model evaluator
Evaluates the legal knowledge of large language models using a custom benchmarking framework.
Benchmarking Legal Knowledge of Large Language Models
273 stars
7 watching
44 forks
Language: Python
last commit: almost 2 years ago benchmarkchatgptlawllm
Related projects:
| Repository | Description | Stars |
|---|---|---|
| | An evaluation toolkit for large vision-language models | 1,514 |
| | Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| | A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. | 566 |
| | A benchmarking framework for large language models | 81 |
| | A Chinese law-focused conversational AI model designed to provide reliable and professional legal answers. | 1,072 |
| | An AI model trained on legal data to provide answers and explanations in Chinese law | 871 |
| | A comprehensive toolkit for evaluating NLP experiments offering automated metrics and efficient computation. | 187 |
| | Develops an NLP pipeline and model for processing long-form legal text | 641 |
| | A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models. | 168 |
| | Evaluates Ontology-Based Data Access systems with inference and meta knowledge benchmarking | 4 |
| | Evaluates language models using standardized benchmarks and prompting techniques. | 2,059 |
| | Develops and trains a large-scale, parameterized model for legal question answering and text generation | 105 |
| | A toolset for evaluating and comparing natural language generation models | 1,350 |
| | An evaluation framework using Chinese high school examination questions to assess large language model capabilities | 565 |
| | A repository of papers and resources for evaluating large language models. | 1,450 |