MuVI

Multi-view modeler

A software framework for multi-view latent variable modeling with domain-informed structured sparsity

A multi-view latent variable model with domain-informed structured sparsity for integrating noisy feature sets.

GitHub

27 stars
5 watching
2 forks
Language: Python
last commit: about 1 month ago
Linked from 1 awesome list


Backlinks from these awesome lists:

Related projects:

Repository Description Stars
tsb0601/mmvp An evaluation framework for multimodal language models' visual capabilities using image and question benchmarks. 296
yuliang-liu/monkey An end-to-end image captioning system that uses large multi-modal models and provides tools for training, inference, and demo usage. 1,849
mcs07/molvs Tool for validating and standardizing chemical structures to improve data quality and facilitate comparisons. 163
vita-mllm/vita A large multimodal language model designed to process and analyze video, image, text, and audio inputs in real-time. 1,005
nvlabs/eagle Develops high-resolution multimodal LLMs by combining vision encoders and various input resolutions 549
yuweihao/mm-vet Evaluates the capabilities of large multimodal models using a set of diverse tasks and metrics 274
opengvlab/multi-modality-arena An evaluation platform for comparing multi-modality models on visual question-answering tasks 478
yfzhang114/slime Develops large multimodal models for high-resolution understanding and analysis of text, images, and other data types. 143
xverse-ai/xverse-v-13b A large multimodal model for visual question answering, trained on a dataset of 2.1B image-text pairs and 8.2M instruction sequences. 78
pku-yuangroup/moe-llava A large vision-language model using a mixture-of-experts architecture to improve performance on multi-modal learning tasks 2,023
zhourax/vega Develops a multimodal task and dataset to assess vision-language models' ability to handle interleaved image-text inputs. 33
freedomintelligence/mllm-bench Evaluates and compares the performance of multimodal large language models on various tasks 56
subho406/omninet An implementation of a unified architecture for multi-modal multi-task learning using PyTorch. 515
chenllliang/mmevalpro A benchmarking framework for evaluating Large Multimodal Models by providing rigorous metrics and an efficient evaluation pipeline. 22
mlpc-ucsd/bliva A multimodal LLM designed to handle text-rich visual questions 270