How To Backdoor Federated Learning
@article{Bagdasaryan2018HowTB,
title={How To Backdoor Federated Learning},
author={Eugene Bagdasaryan and Andreas Veit and Yiqing Hua and Deborah Estrin and Vitaly Shmatikov},
journal={ArXiv},
year={2018},
volume={abs/1807.00459}
}Federated learning enables multiple participants to jointly construct a deep learning model without sharing their private training data with each other. [] Key Result We also show how to evade anomaly detection-based defenses by incorporating the evasion into the loss function when training the attack model.
Figures and Tables from this paper
972 Citations
Backdoor Attacks on Federated Meta-Learning
- 2020
Computer Science
ArXiv
This paper proposes a defense mechanism inspired by matching networks, where the class of an input is predicted from the cosine similarity of its features with a support set of labeled examples, and removes the decision logic from the model shared with the federation.
Defending against Poisoning Backdoor Attacks on Federated Meta-learning
- 2022
Computer Science
ACM Trans. Intell. Syst. Technol.
This article proposes a defense mechanism inspired by matching networks, where the class of an input is predicted from the similarity of its features with a support set of labeled examples, and removes the decision logic from the model shared with the federation.
A Federated Learning Backdoor Attack Defense
- 2022
Computer Science
2022 IEEE Eighth International Conference on Big Data Computing Service and Applications (BigDataService)
A federated learning backdoor attack defense based on dual attention mechanism (FDDAM) is proposed, where the model weights are dynamically adjusted during training process, no additional models are required, and the calculation time is shorter.
DeepSight: Mitigating Backdoor Attacks in Federated Learning Through Deep Model Inspection
- 2022
Computer Science
NDSS
The performance and effectiveness of DeepSight is evaluated and it is shown that it can mitigate state-of-the-art backdoor attacks with a negligible impact on the model’s performance on benign data.
On the Vulnerability of Backdoor Defenses for Federated Learning
- 2023
Computer Science
ArXiv
This paper studies whether the current defense mechanisms truly neutralize the backdoor threats from federated learning in a practical setting by proposing a new federated backdoor attack method for possible countermeasures and provides suggestions to the practitioners when training federated models in practice.
Toward Cleansing Backdoored Neural Networks in Federated Learning
- 2022
Computer Science
2022 IEEE 42nd International Conference on Distributed Computing Systems (ICDCS)
A new and effective method to mitigate backdoor attacks in federated learning after the training phase is proposed, which removes redundant neurons and "backdoor neurons", which trigger misbehavior upon recognizing backdoor patterns while keeping silent when the input data is clean.
Mitigating Backdoor Attacks in Federated Learning
- 2020
Computer Science
ArXiv
A federated pruning method to remove redundant neurons in the network and then adjust the model's extreme weight values is designed, which can reduce the average attack success rate and minimize the pruning influence on test accuracy.
FLAME: Taming Backdoors in Federated Learning
- 2022
Computer Science
USENIX Security Symposium
Evaluation of FLAME on several datasets stemming from application areas including image classification, word prediction, and IoT intrusion detection demonstrates that FLAME removes backdoors effectively with a negligible impact on the benign performance of the models.
Dynamic backdoor attacks against federated learning
- 2020
Computer Science
ArXiv
This paper bridges meta-learning and backdoor attacks under FL setting, in which case the algorithm can learn a versatile model from previous experiences, and fast adapting to new adversarial tasks with a few of examples.
Backdoor Attacks and Defenses in Federated Learning: State-of-the-Art, Taxonomy, and Future Directions
- 2023
Computer Science
IEEE Wireless Communications
This work presents a comprehensive review of the state-of-the-art backdoor attacks and defenses in federated learning, and classifies the existing backdoor attacks into two categories: data poisoning attacks and model poisoning attacks, and divides the defenses into anomaly updates detection, robust federated training, and backdoored model restoration.
80 References
Mitigating Sybils in Federated Learning Poisoning
- 2018
Computer Science
ArXiv
FoolsGold is described, a novel defense to this problem that identifies poisoning sybils based on the diversity of client updates in the distributed learning process that exceeds the capabilities of existing state of the art approaches to countering sybil-based label-flipping and backdoor poisoning attacks.
Analyzing Federated Learning through an Adversarial Lens
- 2019
Computer Science
ICML
This work explores the threat of model poisoning attacks on federated learning initiated by a single, non-colluding malicious agent where the adversarial objective is to cause the model to misclassify a set of chosen inputs with high confidence.
Exploiting Unintended Feature Leakage in Collaborative Learning
- 2019
Computer Science
2019 IEEE Symposium on Security and Privacy (SP)
This work shows that an adversarial participant can infer the presence of exact data points -- for example, specific locations -- in others' training data and develops passive and active inference attacks to exploit this leakage.
BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain
- 2017
Computer Science
ArXiv
It is shown that outsourced training introduces new security risks: an adversary can create a maliciously trained network (a backdoored neural network, or a BadNet) that has state-of-the-art performance on the user's training and validation samples, but behaves badly on specific attacker-chosen inputs.
Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning
- 2017
Computer Science
ArXiv
This work considers a new type of attacks, called backdoor attacks, where the attacker's goal is to create a backdoor into a learning-based authentication system, so that he can easily circumvent the system by leveraging the backdoor.
Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering
- 2019
Computer Science
SafeAI@AAAI
This work proposes a novel approach to backdoor detection and removal for neural networks that is the first methodology capable of detecting poisonous data crafted to insert backdoors and repairing the model that does not require a verified and trusted dataset.
Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks
- 2018
Computer Science
ArXiv
It is shown that even well-generalized models are significantly susceptible to white-box membership inference attacks, by analyzing state-of-the-art pre-trained and publicly available models for the CIFAR dataset.
Auror: defending against poisoning attacks in collaborative deep learning systems
- 2016
Computer Science
ACSAC
This paper investigates the setting of indirect collaborative deep learning --- a form of practical deep learning wherein users submit masked features rather than direct data, and proposes Auror, a system that detects malicious users and generates an accurate model.
Bypassing Backdoor Detection Algorithms in Deep Learning
- 2020
Computer Science
2020 IEEE European Symposium on Security and Privacy (EuroS&P)
An adversarial backdoor embedding algorithm that can bypass the existing detection algorithms including the state-of-the-art techniques, and an adaptive adversarial training algorithm that optimizes the original loss function of the model, and maximizes the indistinguishability of the hidden representations of poisoned data and clean data.
A Little Is Enough: Circumventing Defenses For Distributed Learning
- 2019
Computer Science
NeurIPS
It is shown that 20% of corrupt workers are sufficient to degrade a CIFAR10 model accuracy by 50%, as well as to introduce backdoors into MNIST and CIFar10 models without hurting their accuracy.












