CValues
Model value alignment
Evaluates and aligns the values of Chinese large language models with safety and responsibility standards
面向中文大模型价值观的评估与对齐研究
481 stars
1 watching
20 forks
Language: Python
last commit: over 1 year ago benchmarkchinese-llmsevaluationhuman-valuesllmsmulti-choiceresponsibilitysafety
Related projects:
Repository | Description | Stars |
---|---|---|
| Improves safety and helpfulness of large language models by fine-tuning them using safety-critical tasks | 47 |
| Aligns large language models' behavior through fine-grained correctional human feedback to improve trustworthiness and accuracy. | 245 |
| Measures the understanding of massive multitask Chinese datasets using large language models | 87 |
| Aligns large multimodal models with human intentions and values using various algorithms and fine-tuning methods. | 270 |
| Provides tools to understand and interpret the decisions made by XGBoost models in machine learning | 253 |
| Trains a large Chinese language model on massive data and provides a pre-trained model for downstream tasks | 230 |
| A benchmark for evaluating large language models' ability to process multimodal input | 322 |
| An eXplainability toolbox for machine learning that enables data analysis and model evaluation to mitigate biases and improve performance | 1,135 |
| This project demonstrates the effectiveness of reinforcement learning from human feedback (RLHF) in improving small language models like GPT-2. | 214 |
| An evaluation suite for assessing chart understanding in multimodal large language models. | 85 |
| Evaluates and mitigates hallucinations in multimodal large language models | 82 |
| Generates a benchmark dataset for evaluating protein alignment programs | 3 |
| An MLLM architecture designed to align visual and textual embeddings through structural alignment | 575 |
| Evaluates the capabilities of large multimodal models using a set of diverse tasks and metrics | 274 |
| An implementation of a large language model for Chinese text processing, focusing on MoE (Multi-Headed Attention) architecture and incorporating a vast vocabulary. | 645 |