VL-ICL

Learning benchmark

A benchmarking suite for multimodal in-context learning models

Code for paper: VL-ICL Bench: The Devil in the Details of Benchmarking Multimodal In-Context Learning

GitHub

28 stars
1 watching
2 forks
Language: Python
last commit: 8 months ago

Related projects:

Repository Description Stars
mshukor/evalign-icl Evaluating and improving large multimodal models through in-context learning 20
uw-madison-lee-lab/cobsat Provides a benchmarking framework and dataset for evaluating the performance of large language models in text-to-image tasks 28
ys-zong/vlguard Improves safety and helpfulness of large language models by fine-tuning them using safety-critical tasks 45
ailab-cvc/seed-bench A benchmark for evaluating large language models' ability to process multimodal input 315
yg-smile/rl_vvc_dataset A collection of benchmarks and implementations for testing reinforcement learning-based Volt-VAR control algorithms 20
haozhezhao/mic Develops a multimodal vision-language model to enable machines to understand complex relationships between instructions and images in various tasks. 334
yuliang-liu/multimodalocr An evaluation benchmark for OCR capabilities in large multmodal models. 471
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 14
ydli-ai/csl A large-scale dataset for natural language processing tasks focused on Chinese scientific literature, providing tools and benchmarks for NLP research. 568
cloud-cv/evalai A platform for comparing and evaluating AI and machine learning algorithms at scale 1,771
rll-research/url_benchmark A benchmark suite for unsupervised reinforcement learning agents, providing pre-trained models and scripts for testing and fine-tuning agent performance. 332
scicloj/scicloj.ml.clj-djl Provides pre-trained machine learning models for natural language processing tasks using Clojure and the clj-djl framework. 0
yiren-jian/blitext Develops and trains models for vision-language learning with decoupled language pre-training 24
jiutian-vl/jiutian-lion This project integrates visual knowledge into large language models to improve their capabilities and reduce hallucinations. 121
pku-yuangroup/languagebind Extending pretraining models to handle multiple modalities by aligning language and video representations 723