surrogate_model_extension
Federated Learning Vulnerability Framework
A framework for analyzing and exploiting vulnerabilities in federated learning models using surrogate model attacks
Surrogate Model Extension (SME): A Fast and Accurate Weight Update Attack on Federated Learning [Accepted at ICML 2023]
9 stars
2 watching
1 forks
Language: Python
last commit: 8 months ago Related projects:
Repository | Description | Stars |
---|---|---|
eth-sri/bayes-framework-leakage | Develops and evaluates a framework for detecting attacks on federated learning systems | 11 |
ksreenivasan/ood_federated_learning | Researchers investigate vulnerabilities in Federated Learning systems by introducing new backdoor attacks and exploring methods to defend against them. | 64 |
hfzhang31/a3fl | A framework for attacking federated learning systems with adaptive backdoor attacks | 22 |
jonasgeiping/breaching | A PyTorch framework for analyzing vulnerabilities in federated learning models and predicting data breaches | 269 |
jeremy313/soteria | An implementation of a defense against model inversion attacks in federated learning | 55 |
ai-secure/crfl | This project presents a framework for robust federated learning against backdoor attacks. | 71 |
kenziyuliu/private-cross-silo-fl | This repository provides an implementation of a cross-silo federated learning framework with differential privacy mechanisms. | 25 |
sliencerx/learning-to-attack-federated-learning | An implementation of a framework for learning how to attack federated learning systems | 15 |
yunqing-me/attackvlm | An adversarial attack framework on large vision-language models | 161 |
ai-secure/dba | A tool for demonstrating and analyzing attacks on federated learning systems by introducing backdoors into distributed machine learning models. | 176 |
jeremy313/fl-wbc | A defense mechanism against model poisoning attacks in federated learning | 37 |
shenzebang/centaur-privacy-federated-representation-learning | A framework for Federated Learning with Differential Privacy using PyTorch | 13 |
inspire-group/modelpoisoning | An implementation of model poisoning attacks in federated learning | 146 |
fangxiuwen/robust_fl | An implementation of a robust federated learning framework for handling noisy and heterogeneous clients in machine learning. | 41 |
zlijingtao/ressfl | Develops techniques to improve the resistance of split learning in federated learning against model inversion attacks | 20 |