SEED-Bench
Multimodal LLM test suite
A benchmark for evaluating large language models' ability to process multimodal input
(CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.
322 stars
4 watching
13 forks
Language: Python
last commit: 8 months ago Related projects:
Repository | Description | Stars |
---|---|---|
| An implementation of a multimodal language model with capabilities for comprehension and generation | 585 |
| Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. | 15 |
| An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models | 84 |
| Evaluating and improving large multimodal models through in-context learning | 21 |
| Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| An implementation of a multimodal language model using locality-enhanced projection techniques | 435 |
| An evaluation benchmark for OCR capabilities in large multmodal models. | 484 |
| Automated machine learning protocols for cheminformatics using Python | 39 |
| A benchmark for evaluating large language models in multiple languages and formats | 93 |
| A benchmarking suite for multimodal in-context learning models | 31 |
| A multimodal LLM designed to handle text-rich visual questions | 270 |
| Develops a large-scale dataset and benchmark for training multimodal chart understanding models using large language models. | 87 |
| A benchmarking framework for large language models | 81 |
| A platform for comparing and evaluating AI and machine learning algorithms at scale | 1,779 |
| An LLM-free benchmark suite for evaluating MLLMs' hallucination capabilities in various tasks and dimensions | 98 |