SEED-Bench

Multimodal LLM test suite

A benchmark for evaluating large language models' ability to process multimodal input

(CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.

GitHub

322 stars
4 watching
13 forks
Language: Python
last commit: 5 months ago

Related projects:

Repository Description Stars
ailab-cvc/seed An implementation of a multimodal language model with capabilities for comprehension and generation 585
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 15
aifeg/benchlmm An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models 84
mshukor/evalign-icl Evaluating and improving large multimodal models through in-context learning 21
freedomintelligence/mllm-bench Evaluates and compares the performance of multimodal large language models on various tasks 56
khanrc/honeybee An implementation of a multimodal language model using locality-enhanced projection techniques 435
yuliang-liu/multimodalocr An evaluation benchmark for OCR capabilities in large multmodal models. 484
jvalegre/robert Automated machine learning protocols for cheminformatics using Python 39
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 93
ys-zong/vl-icl A benchmarking suite for multimodal in-context learning models 31
mlpc-ucsd/bliva A multimodal LLM designed to handle text-rich visual questions 270
fuxiaoliu/mmc Develops a large-scale dataset and benchmark for training multimodal chart understanding models using large language models. 87
qcri/llmebench A benchmarking framework for large language models 81
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,779
junyangwang0410/amber An LLM-free benchmark suite for evaluating MLLMs' hallucination capabilities in various tasks and dimensions 98