VLGuard
Safety fine-tuner
Improves safety and helpfulness of large language models by fine-tuning them using safety-critical tasks
[ICML 2024] Safety Fine-Tuning at (Almost) No Cost: A Baseline for Vision Large Language Models.
47 stars
3 watching
2 forks
Language: Python
last commit: 6 months ago alignmentlarge-language-modelslarge-vision-language-modelssafetyvision-language-model
Related projects:
Repository | Description | Stars |
---|---|---|
| A tool to streamline fine-tuning of multimodal models for vision-language tasks | 1,415 |
| A benchmarking suite for multimodal in-context learning models | 31 |
| A tool for generating and evaluating multimodal Large Language Models with visual instruction tuning capabilities | 93 |
| Evaluates and aligns the values of Chinese large language models with safety and responsibility standards | 481 |
| Aligns large language models' behavior through fine-grained correctional human feedback to improve trustworthiness and accuracy. | 245 |
| A benchmark for evaluating the safety and robustness of vision language models against adversarial attacks. | 72 |
| Debiasing techniques to minimize hallucinations in large visual language models | 75 |
| A comprehensive dataset and evaluation framework for Vision-Language Instruction Tuning models | 11 |
| Improves pre-trained Chinese language models by incorporating a correction task to alleviate inconsistency issues with downstream tasks | 646 |
| Analyzing and mitigating object hallucination in large vision-language models to improve their accuracy and reliability. | 136 |
| Provides guidance on fine-tuning pre-trained models for image classification tasks using PyTorch. | 279 |
| A unified benchmark for safe reinforcement learning algorithms and environments. | 410 |
| Extending pretraining models to handle multiple modalities by aligning language and video representations | 751 |
| A Chinese finance-focused large language model fine-tuning framework | 596 |
| Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |