TrustLLM

Trust Assessment Toolkit

A toolkit for assessing trustworthiness in large language models

[ICML 2024] TrustLLM: Trustworthiness in Large Language Models

GitHub

491 stars
8 watching
47 forks
Language: Python
last commit: 4 months ago
Linked from 1 awesome list

aibenchmarkdatasetevaluationlarge-language-modelsllmnatural-language-processingnlppypi-packagetoolkittrustworthy-aitrustworthy-machine-learning

Backlinks from these awesome lists:

Related projects:

Repository Description Stars
aiplanethub/beyondllm An open-source toolkit for building and evaluating large language models 267
vhellendoorn/code-lms A guide to using pre-trained large language models in source code analysis and generation 1,789
leondz/lm_risk_cards A set of tools and guidelines for assessing the security vulnerabilities of language models in AI applications 28
mlgroupjlu/llm-eval-survey A repository of papers and resources for evaluating large language models. 1,450
nlpai-lab/kullm Korea University Large Language Model developed by researchers at Korea University and HIAI Research Institute. 576
academic-hammer/hammerllm A large language model pre-trained on Chinese and English data, suitable for natural language processing tasks. 43
safellama/plexiglass A toolkit to detect and protect against vulnerabilities in Large Language Models. 122
damo-nlp-sg/m3exam A benchmark for evaluating large language models in multiple languages and formats 93
phodal/aigc Developing and applying large language models to improve software development workflows and processes 1,413
csuhan/onellm A framework for training and fine-tuning multimodal language models on various data types 601
trusted-ai/aix360 A toolkit for explaining complex AI models and data-driven insights 1,641
ucsc-vlaa/vllm-safety-benchmark A benchmark for evaluating the safety and robustness of vision language models against adversarial attacks. 72
michael-wzhu/shennong-tcm-llm Develops and deploys a large language model for Chinese traditional medicine applications 316
luogen1996/lavin An open-source implementation of a vision-language instructed large language model 513
aifeg/benchlmm An open-source benchmarking framework for evaluating cross-style visual capability of large multimodal models 84