Visual-CoT

Chain-of-Thought Model

Develops a multi-modal language model with a comprehensive dataset and benchmark for chain-of-thought reasoning

[Neurips'24 Spotlight] Visual CoT: Advancing Multi-Modal Language Models with a Comprehensive Dataset and Benchmark for Chain-of-Thought Reasoning

GitHub

134 stars
1 watching
7 forks
Language: Python
last commit: about 1 month ago

Related projects:

Repository Description Stars
rowanz/r2c An open-source project providing PyTorch code and data for a deep learning model that enables visual commonsense reasoning. 466
lxtgh/omg-seg Develops an end-to-end model for multiple visual perception and reasoning tasks using a single encoder, decoder, and large language model. 1,300
zhegan27/semantic_compositional_nets A deep learning framework providing a model architecture and training code for image captioning using semantic compositional networks 70
davidmascharka/tbd-nets An open-source implementation of a deep learning model designed to improve the balance between performance and interpretability in visual reasoning tasks. 348
cod3licious/conec A library for training and evaluating a type of word embedding model that extends the original Word2Vec algorithm 20
rucaibox/comvint Creating synthetic visual reasoning instructions to improve the performance of large language models on image-related tasks 18
cadene/vqa.pytorch A PyTorch implementation of visual question answering with multimodal representation learning 716
bigredt/vico Multi-sense word embeddings learned from visual cooccurrences 25
360cvgroup/360vl A large multi-modal model developed using the Llama3 language model, designed to improve image understanding capabilities. 30
deepseek-ai/deepseek-vl A multimodal AI model that enables real-world vision-language understanding applications 2,077
tianyi-lab/hallusionbench An image-context reasoning benchmark designed to challenge large vision-language models and help improve their accuracy 243
kazuto1011/deeplab-pytorch PyTorch implementation of DeepLab v2 for semantic segmentation on COCO-Stuff and PASCAL VOC datasets 1,093
jayleicn/clipbert An efficient framework for end-to-end learning on image-text and video-text tasks 704
satwikkottur/visualword2vec Learning word embeddings from abstract images to improve language understanding 19
kdexd/virtex A pretraining approach that uses semantically dense captions to learn visual representations and improve image understanding tasks. 557