MIC

Multimodal learner

Develops a multimodal vision-language model to enable machines to understand complex relationships between instructions and images in various tasks.

MMICL, a state-of-the-art VLM with the in context learning ability from ICL, PKU

GitHub

334 stars
10 watching
15 forks
Language: Python
last commit: 11 months ago

Related projects:

Repository Description Stars
fuxiaoliu/mmc Develops a large-scale dataset and benchmark for training multimodal chart understanding models using large language models. 84
pleisto/yuren-baichuan-7b A multi-modal large language model that integrates natural language and visual capabilities with fine-tuning for various tasks 72
ys-zong/vl-icl A benchmarking suite for multimodal in-context learning models 28
lyuchenyang/macaw-llm A multi-modal language model that integrates image, video, audio, and text data to improve language understanding and generation 1,550
byungkwanlee/moai Improves performance of vision language tasks by integrating computer vision capabilities into large language models 311
pku-yuangroup/languagebind Extending pretraining models to handle multiple modalities by aligning language and video representations 723
yuliang-liu/monkey A toolkit for building conversational AI models that can process images and text inputs. 1,825
yuweihao/mm-vet Evaluates the capabilities of large multimodal models using a set of diverse tasks and metrics 267
mbzuai-oryx/groundinglmm An end-to-end trained model capable of generating natural language responses integrated with object segmentation masks. 781
pku-yuangroup/moe-llava Develops a neural network architecture for multi-modal learning with large vision-language models 1,980
mshukor/evalign-icl Evaluating and improving large multimodal models through in-context learning 20
zhourax/vega Develops a multimodal task and dataset to assess vision-language models' ability to handle interleaved image-text inputs. 33
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 14
jshilong/gpt4roi Training and deploying large language models on computer vision tasks using region-of-interest inputs 506
open-mmlab/multimodal-gpt Trains a multimodal chatbot that combines visual and language instructions to generate responses 1,477