Federated-Learning-Backdoor
Federated Learning Attack
An implementation of a federated learning attack method known as Neurotoxin, which introduces backdoors into machine learning models during the training process.
ICML 2022 code for "Neurotoxin: Durable Backdoors in Federated Learning" https://arxiv.org/abs/2206.10341
63 stars
3 watching
7 forks
Language: Python
last commit: over 1 year ago backdoor-attacksfederated-learning
Related projects:
Repository | Description | Stars |
---|---|---|
hfzhang31/a3fl | A framework for attacking federated learning systems with adaptive backdoor attacks | 22 |
ebagdasa/backdoor_federated_learning | An implementation of a framework for backdoors in federated learning, allowing researchers to test and analyze various attacks on distributed machine learning models. | 271 |
ai-secure/dba | A tool for demonstrating and analyzing attacks on federated learning systems by introducing backdoors into distributed machine learning models. | 176 |
git-disl/lockdown | A backdoor defense system against attacks in federated learning algorithms used for machine learning model training on distributed datasets. | 14 |
ai-secure/crfl | This project presents a framework for robust federated learning against backdoor attacks. | 71 |
ybdai7/chameleon-durable-backdoor | A federated learning system implementation that enables planting durable backdoors in global models by adapting to peer images. | 32 |
dcalab-unipv/turning-privacy-preserving-mechanisms-against-federated-learning | This project presents an attack on federated learning systems to compromise their privacy-preserving mechanisms. | 8 |
ksreenivasan/ood_federated_learning | Researchers investigate vulnerabilities in Federated Learning systems by introducing new backdoor attacks and exploring methods to defend against them. | 64 |
jeremy313/fl-wbc | A defense mechanism against model poisoning attacks in federated learning | 37 |
ai-secure/fedgame | An implementation of a game-theoretic defense against backdoor attacks in federated learning. | 5 |
jeremy313/soteria | An implementation of a defense against model inversion attacks in federated learning | 55 |
jonasgeiping/breaching | A PyTorch framework for analyzing vulnerabilities in federated learning models and predicting data breaches | 269 |
zhuohangli/ggl | An attack implementation to test and evaluate the effectiveness of federated learning privacy defenses. | 57 |
sliencerx/learning-to-attack-federated-learning | An implementation of a framework for learning how to attack federated learning systems | 15 |
deu30303/feddefender | A PyTorch implementation of an attack-tolerant federated learning system to train robust local models against malicious attacks from adversaries. | 9 |