align-anything

Model aligner

Aligns large multimodal models with human intentions and values using various algorithms and fine-tuning methods.

Align Anything: Training All-modality Model with Feedback

GitHub

270 stars
9 watching
53 forks
Language: Python
last commit: about 1 month ago
chameleondpolarge-language-modelsmultimodalrlhfvision-language-model

Related projects:

Repository Description Stars
pku-yuangroup/languagebind Extending pretraining models to handle multiple modalities by aligning language and video representations 751
aidc-ai/ovis An MLLM architecture designed to align visual and textual embeddings through structural alignment 575
ethanyanjiali/minchatgpt This project demonstrates the effectiveness of reinforcement learning from human feedback (RLHF) in improving small language models like GPT-2. 214
lancopku/iais This project proposes a novel method for calibrating attention distributions in multimodal models to improve contextualized representations of image-text pairs. 30
x-plug/cvalues Evaluates and aligns the values of Chinese large language models with safety and responsibility standards 481
rlhf-v/rlhf-v Aligns large language models' behavior through fine-grained correctional human feedback to improve trustworthiness and accuracy. 245
pku-yuangroup/video-bench Evaluates and benchmarks large language models' video understanding capabilities 121
pkunlp-icler/pca-eval An open-source benchmark and evaluation tool for assessing multimodal large language models' performance in embodied decision-making tasks 99
cmesher/inuktitutalignerdata Scripts for aligning laboratory speech production data in Inuktitut 3
pku-yuangroup/moe-llava A large vision-language model using a mixture-of-experts architecture to improve performance on multi-modal learning tasks 2,023
prosodylab/prosodylab-aligner Tools for aligning laboratory speech production data to forced audio alignment using HTK and SoX. 333
mshukor/evalign-icl Evaluating and improving large multimodal models through in-context learning 21
multimodal-art-projection/omnibench Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. 15
jcgood/rosetta-pangloss A Python library that uses machine learning and natural language processing to improve translation accuracy by aligning source and target languages 0
opengvlab/multi-modality-arena An evaluation platform for comparing multi-modality models on visual question-answering tasks 478