jailbreak-evaluation
Control evaluation
Evaluates language model attempts to determine their control and trustworthiness
The jailbreak-evaluation is an easy-to-use Python package for language model jailbreak evaluation.
20 stars
0 watching
3 forks
Language: Python
last commit: 4 months ago
Linked from 1 awesome list
Related projects:
Repository | Description | Stars |
---|---|---|
| Develops an interpretable evaluation procedure for off-policy evaluation (OPE) methods to quantify their sensitivity to hyper-parameter choices and/or evaluation policy choices. | 31 |
| Evaluates and compares the performance of multimodal large language models on various tasks | 56 |
| An evaluation suite and dynamic data release platform for large language models | 230 |
| A framework for evaluating and monitoring the quality of large language model outputs in Retrieval Augmented Generation applications. | 271 |
| A Python library designed to support the development of timing-critical experiments in cognitive science and neuroscience. | 115 |
| Provides a way to run Behavior-Driven Development tests in Django applications | 197 |
| Evaluates language models using standardized benchmarks and prompting techniques. | 2,059 |
| A comprehensive Python toolbox for evaluating salient object detection and camouflaged object detection tasks | 168 |
| An evaluation framework for large language models trained with instruction tuning methods | 535 |
| An evaluation suite for assessing chart understanding in multimodal large language models. | 85 |
| A BDD testing framework for Django applications | 205 |
| Improves safety and helpfulness of large language models by fine-tuning them using safety-critical tasks | 47 |
| A framework for evaluating language models on NLP tasks | 326 |
| A framework for evaluating large language models | 669 |
| Tools and resources for identifying biased language in code and content. | 21 |