align-anything

Model aligner

Aligns large models with human values and intentions across various modalities.

Align Anything: Training All-modality Model with Feedback

GitHub

244 stars
8 watching
43 forks
Language: Python
last commit: 11 days ago
chameleondpolarge-language-modelsmultimodalrlhfvision-language-model

Related projects:

Repository Description Stars
pku-yuangroup/languagebind Extending pretraining models to handle multiple modalities by aligning language and video representations 723
aidc-ai/ovis An architecture designed to align visual and textual embeddings in multimodal learning 517
ethanyanjiali/minchatgpt This project demonstrates the effectiveness of reinforcement learning from human feedback (RLHF) in improving small language models like GPT-2. 213
lancopku/iais This project proposes a novel method for calibrating attention distributions in multimodal models to improve contextualized representations of image-text pairs. 30
x-plug/cvalues Evaluates and aligns the values of Chinese large language models with safety and responsibility standards 477
rlhf-v/rlhf-v Aligns large language models' behavior through fine-grained correctional human feedback to improve trustworthiness and accuracy. 233
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 117
pkunlp-icler/pca-eval An open-source benchmark and evaluation tool for assessing multimodal large language models' performance in embodied decision-making tasks 100
cmesher/inuktitutalignerdata Tools for aligning laboratory speech production data 3
pku-yuangroup/moe-llava Develops a neural network architecture for multi-modal learning with large vision-language models 1,980
prosodylab/prosodylab-aligner A Python tool for aligning audio data from laboratory speech production experiments 331
mshukor/evalign-icl Evaluating and improving large multimodal models through in-context learning 20
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 14
jcgood/rosetta-pangloss A Python library that uses machine learning and natural language processing to improve translation accuracy by aligning source and target languages 0
opengvlab/multi-modality-arena An evaluation platform for comparing multi-modality models on visual question-answering tasks 467