Federated-Learning-Backdoor

Federated Learning Attack

An implementation of a federated learning attack method known as Neurotoxin, which introduces backdoors into machine learning models during the training process.

ICML 2022 code for "Neurotoxin: Durable Backdoors in Federated Learning" https://arxiv.org/abs/2206.10341

GitHub

65 stars
3 watching
7 forks
Language: Python
last commit: almost 2 years ago
backdoor-attacksfederated-learning

Related projects:

Repository Description Stars
hfzhang31/a3fl A framework for attacking federated learning systems with adaptive backdoor attacks 23
ebagdasa/backdoor_federated_learning This project provides an implementation of backdoor attacks in federated learning frameworks using Python and PyTorch. 277
ai-secure/dba A tool for demonstrating and analyzing attacks on federated learning systems by introducing backdoors into distributed machine learning models. 179
git-disl/lockdown A backdoor defense system for federated learning, designed to protect against data poisoning attacks by isolating subspace training and aggregating models with robust consensus fusion. 18
ai-secure/crfl This project presents a framework for robust federated learning against backdoor attacks. 71
ybdai7/chameleon-durable-backdoor A federated learning system implementation that enables planting durable backdoors in global models by adapting to peer images. 34
dcalab-unipv/turning-privacy-preserving-mechanisms-against-federated-learning This project presents an attack on federated learning systems to compromise their privacy-preserving mechanisms. 8
ksreenivasan/ood_federated_learning Researchers investigate vulnerabilities in Federated Learning systems by introducing new backdoor attacks and exploring methods to defend against them. 66
jeremy313/fl-wbc A defense mechanism against model poisoning attacks in federated learning 37
ai-secure/fedgame An implementation of a game-theoretic defense against backdoor attacks in federated learning. 6
jeremy313/soteria An implementation of a defense against model inversion attacks in federated learning 55
jonasgeiping/breaching A PyTorch framework for analyzing vulnerabilities in federated learning models and predicting data breaches 274
zhuohangli/ggl Researchers develop an attack method to measure the effectiveness of federated learning privacy defenses by generating leakage in gradients 58
sliencerx/learning-to-attack-federated-learning An implementation of a framework for learning how to attack federated learning systems 15
deu30303/feddefender A PyTorch implementation of an attack-tolerant federated learning system to train robust local models against malicious attacks from adversaries. 10