LawBench
Legal model evaluator
Evaluates the legal knowledge of large language models using a custom benchmarking framework.
Benchmarking Legal Knowledge of Large Language Models
273 stars
7 watching
44 forks
Language: Python
last commit: over 1 year ago benchmarkchatgptlawllm
Related projects:
Repository | Description | Stars |
---|---|---|
| An evaluation toolkit for large vision-language models | 1,514 |
| Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| A tool to automate the evaluation of large language models in Google Colab using various benchmarks and custom parameters. | 566 |
| A benchmarking framework for large language models | 81 |
| A Chinese law-focused conversational AI model designed to provide reliable and professional legal answers. | 1,072 |
| An AI model trained on legal data to provide answers and explanations in Chinese law | 871 |
| A comprehensive toolkit for evaluating NLP experiments offering automated metrics and efficient computation. | 187 |
| Develops an NLP pipeline and model for processing long-form legal text | 641 |
| A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models. | 168 |
| Evaluates Ontology-Based Data Access systems with inference and meta knowledge benchmarking | 4 |
| Evaluates language models using standardized benchmarks and prompting techniques. | 2,059 |
| Develops and trains a large-scale, parameterized model for legal question answering and text generation | 105 |
| A toolset for evaluating and comparing natural language generation models | 1,350 |
| An evaluation framework using Chinese high school examination questions to assess large language model capabilities | 565 |
| A repository of papers and resources for evaluating large language models. | 1,450 |