R-GAP

Gradient attack tool

A tool to demonstrate and analyze attacks on private data in machine learning models using gradients

R-GAP: Recursive Gradient Attack on Privacy [Accepted at ICLR 2021]

GitHub

34 stars
2 watching
2 forks
Language: Python
last commit: almost 2 years ago

Related projects:

Repository Description Stars
superhedgy/attacksurfacemapper Automates reconnaissance to identify potential attack surfaces of a target network 1,314
yunqing-me/attackvlm An adversarial attack framework on large vision-language models 161
patrickzh/improved-deep-leakage-from-gradients This project presents an improved method to extract accurate private training data from publicly shared gradients in distributed learning systems. 144
zhuohangli/ggl An attack implementation to test and evaluate the effectiveness of federated learning privacy defenses. 57
zou-group/textgrad An autograd engine for textual gradients using large language models to backpropagate gradients. 1,821
ml-postech/gradient-inversion-generative-image-prior An implementation of a method to invert gradients in federated learning to potentially reveal sensitive client data 39
ys-zong/foolyourvllms An attack framework to manipulate the output of large language models and vision-language models 14
eth-sri/lamp This is a research project that extracts text from gradients with language model priors using deep learning techniques. 23
johnkorzhuk/grabient A tool to generate linear web gradients with a user interface. 1,990
borealisai/advertorch A toolbox for researching and evaluating robustness against attacks on machine learning models 1,308
junyizhu-ai/surrogate_model_extension A framework for analyzing and exploiting vulnerabilities in federated learning models using surrogate model attacks 9
lhfowl/robbing_the_fed This implementation allows an attacker to directly obtain user data from federated learning gradient updates by modifying the shared model architecture. 23
jind11/textfooler A tool for generating adversarial examples to attack text classification and inference models 494
eth-sri/diffai Trains neural networks to be provably robust against adversarial examples using abstract interpretation techniques. 218
ai4risk/antifraud Develops and evaluates machine learning models for detecting financial fraud 174