FoolyourVLLMs
Attack framework
An attack framework to manipulate the output of large language models and vision-language models
[ICML 2024] Fool Your (Vision and) Language Model With Embarrassingly Simple Permutations
14 stars
1 watching
2 forks
Language: Python
last commit: over 1 year ago adversarial-attacksllmsmcqvision-and-language
Related projects:
Repository | Description | Stars |
---|---|---|
| An adversarial attack framework on large vision-language models | 165 |
| Improves safety and helpfulness of large language models by fine-tuning them using safety-critical tasks | 47 |
| A benchmarking suite for multimodal in-context learning models | 31 |
| A framework for attacking federated learning systems with adaptive backdoor attacks | 23 |
| Detecting backdoors in language models to prevent malicious AI usage | 109 |
| A framework for large-scale cross-modal benchmarks and vision-language tasks in Chinese | 157 |
| A defense mechanism against model poisoning attacks in federated learning | 37 |
| A tool to demonstrate and analyze attacks on private data in machine learning models using gradients | 34 |
| A framework for training and utilizing large language models with knowledge augmentation capabilities | 1,251 |
| An open-source reinforcement learning framework to generate adversarial examples for malware classification models. | 41 |
| This implementation allows an attacker to directly obtain user data from federated learning gradient updates by modifying the shared model architecture. | 23 |
| Analyzing and mitigating object hallucination in large vision-language models to improve their accuracy and reliability. | 136 |
| A framework for defending against backdoor attacks in federated learning systems | 48 |
| An end-to-end image captioning system that uses large multi-modal models and provides tools for training, inference, and demo usage. | 1,849 |
| A tool for generating adversarial examples to attack text classification and inference models | 496 |