PromptInject
Prompt analysis tool
A framework for analyzing the robustness of large language models to adversarial prompt attacks
PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to adversarial prompt attacks. 🏆 Best Paper Awards @ NeurIPS ML Safety Workshop 2022
318 stars
11 watching
32 forks
Language: Python
last commit: 12 months ago adversarial-attacksagiagi-alignmentai-alignmentai-safetychain-of-thoughtgpt-3language-modelslarge-language-modelsmachine-learningml-safetyprompt-engineering
Related projects:
Repository | Description | Stars |
---|---|---|
| Protects AI applications from prompt injection attacks through multiple layers of defense | 1,144 |
| A guide to help developers understand and mitigate the security risks of prompt injection in AI-powered applications and features. | 376 |
| A comprehensive resource for learning prompt engineering techniques for interacting with large language models. | 36 |
| An API for evaluating the quality of text prompts used in Large Language Models (LLMs) based on perplexity estimation | 43 |
| A tool to reduce the complexity of text prompts to minimize API costs and model computations. | 246 |
| A framework to help security analysts understand and prepare for adversarial machine learning attacks on AI systems | 1,056 |
| A unified framework for evaluating large language models' performance and robustness in various scenarios. | 2,487 |
| A research-friendly codebase for experimenting with multi-agent reinforcement learning in JAX | 749 |
| An interactive tool that tests and hardens the security of system prompts used in GenAI applications against various attacks. | 419 |
| An ExpressJS middleware that allows users to execute LLM prompts stored in a git repository and retrieve results from a chosen model. | 74 |
| An approach to leveraging pre-trained models for efficient speech processing tasks by using prompt tuning | 97 |
| A security scanner for Large Language Model prompts to detect potential threats and vulnerabilities | 326 |
| Standardized framework for creating and sharing incident response processes in a shared language | 151 |
| A comprehensive collection of customizable prompts for Generative Pre-trained Transformers (GPTs) designed specifically for educational use. | 77 |
| A Rust implementation of a minimal, fast, and aesthetically pleasing prompt system | 252 |