MMBench
Multi-modal model evaluation suite
A collection of benchmarks to evaluate the multi-modal understanding capability of large vision language models.
Official Repo of "MMBench: Is Your Multi-modal Model an All-around Player?"
168 stars
3 watching
10 forks
last commit: 6 months ago Related projects:
Repository | Description | Stars |
---|---|---|
| An evaluation toolkit for large vision-language models | 1,514 |
| Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. | 15 |
| A family of large multimodal models supporting multimodal conversational capabilities and text-to-image generation in multiple languages | 1,098 |
| Evaluates the legal knowledge of large language models using a custom benchmarking framework. | 273 |
| An evaluation platform for comparing multi-modality models on visual question-answering tasks | 478 |
| An empirical study aiming to develop a large language model capable of effectively integrating multiple input modalities | 23 |
| An open-source software project providing a comprehensive 3D instruction-following dataset with multi-modal prompts for training large language models. | 58 |
| Develops a large-scale dataset and benchmark for training multimodal chart understanding models using large language models. | 87 |
| An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. | 296 |
| A benchmarking framework designed to evaluate the robustness of large multimodal models against common corruption scenarios | 27 |
| A MATLAB package for modeling and analyzing multivariate neural responses to dynamic stimuli. | 85 |
| Provides a modular framework and tools for working with 3D human parametric models in computer vision and graphics | 1,253 |
| An open-source toolbox for action understanding from video data using PyTorch. | 1,863 |
| A benchmark for evaluating large language models' ability to process multimodal input | 322 |
| A multi-modal large language model that integrates natural language and visual capabilities with fine-tuning for various tasks | 73 |