align-anything
Model aligner
Aligns large models with human values and intentions across various modalities.
Align Anything: Training All-modality Model with Feedback
244 stars
8 watching
43 forks
Language: Python
last commit: 11 days ago chameleondpolarge-language-modelsmultimodalrlhfvision-language-model
Related projects:
Repository | Description | Stars |
---|---|---|
pku-yuangroup/languagebind | Extending pretraining models to handle multiple modalities by aligning language and video representations | 723 |
aidc-ai/ovis | An architecture designed to align visual and textual embeddings in multimodal learning | 517 |
ethanyanjiali/minchatgpt | This project demonstrates the effectiveness of reinforcement learning from human feedback (RLHF) in improving small language models like GPT-2. | 213 |
lancopku/iais | This project proposes a novel method for calibrating attention distributions in multimodal models to improve contextualized representations of image-text pairs. | 30 |
x-plug/cvalues | Evaluates and aligns the values of Chinese large language models with safety and responsibility standards | 477 |
rlhf-v/rlhf-v | Aligns large language models' behavior through fine-grained correctional human feedback to improve trustworthiness and accuracy. | 233 |
pku-yuangroup/video-bench | Evaluates and benchmarks large language models' video understanding capabilities | 117 |
pkunlp-icler/pca-eval | An open-source benchmark and evaluation tool for assessing multimodal large language models' performance in embodied decision-making tasks | 100 |
cmesher/inuktitutalignerdata | Tools for aligning laboratory speech production data | 3 |
pku-yuangroup/moe-llava | Develops a neural network architecture for multi-modal learning with large vision-language models | 1,980 |
prosodylab/prosodylab-aligner | A Python tool for aligning audio data from laboratory speech production experiments | 331 |
mshukor/evalign-icl | Evaluating and improving large multimodal models through in-context learning | 20 |
multimodal-art-projection/omnibench | Evaluates and benchmarks multimodal language models' ability to process visual, acoustic, and textual inputs simultaneously. | 14 |
jcgood/rosetta-pangloss | A Python library that uses machine learning and natural language processing to improve translation accuracy by aligning source and target languages | 0 |
opengvlab/multi-modality-arena | An evaluation platform for comparing multi-modality models on visual question-answering tasks | 467 |