TrustLLM

Trust Assessment Toolkit

A toolkit for assessing trustworthiness in large language models

[ICML 2024] TrustLLM: Trustworthiness in Large Language Models

GitHub

466 stars
8 watching
44 forks
Language: Python
last commit: about 2 months ago
Linked from 1 awesome list

aibenchmarkdatasetevaluationlarge-language-modelsllmnatural-language-processingnlppypi-packagetoolkittrustworthy-aitrustworthy-machine-learning

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
aiplanethub/beyondllm An open-source toolkit for building and evaluating large language models 263
vhellendoorn/code-lms A guide to using pre-trained large language models in source code analysis and generation 1,782
leondz/lm_risk_cards A set of tools and guidelines for assessing the security vulnerabilities of language models in AI applications 25
mlgroupjlu/llm-eval-survey A repository of papers and resources for evaluating large language models. 1,433
nlpai-lab/kullm Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. 569
academic-hammer/hammerllm A large language model pre-trained on Chinese and English data, suitable for natural language processing tasks. 43
safellama/plexiglass A toolkit to detect and protect against vulnerabilities in Large Language Models. 121
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 92
phodal/aigc Developing and applying large language models to improve software development workflows and processes 1,386
csuhan/onellm A framework for training and fine-tuning multimodal language models on various data types 588
trusted-ai/aix360 A toolkit for explaining complex AI models and data-driven insights 1,633
ucsc-vlaa/vllm-safety-benchmark A benchmark for evaluating the safety and robustness of vision language models against adversarial attacks. 67
michael-wzhu/shennong-tcm-llm Develops and deploys a large language model for Chinese traditional medicine applications 299
luogen1996/lavin An open-source implementation of a vision-language instructed large language model 508
aifeg/benchlmm An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models 82