Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning
- Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, D. Song
- Computer ScienceArXiv
- 15 December 2017
This work considers a new type of attacks, called backdoor attacks, where the attacker's goal is to create a backdoor into a learning-based authentication system, so that he can easily circumvent the system by leveraging the backdoor.
Robust Physical-World Attacks on Deep Learning Visual Classification
- Kevin Eykholt, I. Evtimov, D. Song
- Computer ScienceIEEE/CVF Conference on Computer Vision and…
- 1 June 2018
This work proposes a general attack algorithm, Robust Physical Perturbations (RP2), to generate robust visual adversarial perturbations under different physical conditions and shows that adversarial examples generated using RP2 achieve high targeted misclassification rates against standard-architecture road sign classifiers in the physical world under various environmental conditions, including viewpoints.
Towards Stable and Efficient Training of Verifiably Robust Neural Networks
- Huan Zhang, Hongge Chen, Chaowei Xiao, Bo Li, D. Boning, Cho-Jui Hsieh
- Computer ScienceInternational Conference on Learning…
- 14 June 2019
CROWN-IBP is computationally efficient and consistently outperforms IBP baselines on training verifiably robust neural networks, and outperform all previous linear relaxation and bound propagation based certified defenses in $\ell_\infty$ robustness.
Generating Adversarial Examples with Adversarial Networks
- Chaowei Xiao, Bo Li, Jun-Yan Zhu, Warren He, M. Liu, D. Song
- Computer ScienceInternational Joint Conference on Artificial…
- 8 January 2018
Adversarial examples generated by AdvGAN on different target models have high attack success rate under state-of-the-art defenses compared to other attacks, and have placed the first with 92.76% accuracy on a public MNIST black-box attack challenge.
Spatially Transformed Adversarial Examples
- Chaowei Xiao, Jun-Yan Zhu, Bo Li, Warren He, M. Liu, D. Song
- Computer ScienceInternational Conference on Learning…
- 8 January 2018
Perturbations generated through spatial transformation could result in large $\mathcal{L}_p$ distance measures, but the extensive experiments show that such spatially transformed adversarial examples are perceptually realistic and more difficult to defend against with existing defense systems.
Neural Attention Distillation: Erasing Backdoor Triggers from Deep Neural Networks
- Yige Li, Nodens Koren, L. Lyu, X. Lyu, Bo Li, Xingjun Ma
- Computer ScienceInternational Conference on Learning…
- 15 January 2021
This paper proposes a novel defense framework Neural Attention Distillation (NAD), which utilizes a teacher network to guide the finetuning of the backdoored student network on a small clean subset of data such that the intermediate-layer attention of the student network aligns with that of the teacher network.
Physical Adversarial Examples for Object Detectors
- Kevin Eykholt, I. Evtimov, D. Song
- Computer ScienceWOOT @ USENIX Security Symposium
- 20 July 2018
This work improves upon a previous physical attack on image classifiers, and creates perturbed physical objects that are either ignored or mislabeled by object detection models, and implements a Disappearance Attack, which causes a Stop sign to "disappear" according to the detector.
Characterizing Audio Adversarial Examples Using Temporal Dependency
- Zhuolin Yang, Bo Li, Pin-Yu Chen, D. Song
- Computer ScienceInternational Conference on Learning…
- 27 September 2018
The results reveal the importance of using the temporal dependency in audio data to gain discriminate power against adversarial examples and offer novel insights in exploiting domain-specific data properties to mitigate negative effects of adversarialExamples.
Robust Physical-World Attacks on Deep Learning Models
- I. Evtimov, Kevin Eykholt, D. Song
- Computer Science
- 27 July 2017
This work proposes a general attack algorithm,Robust Physical Perturbations (RP2), to generate robust visual adversarial perturbations under different physical conditions and shows that adversarial examples generated using RP2 achieve high targeted misclassification rates against standard-architecture road sign classifiers in the physical world under various environmental conditions, including viewpoints.
Robust Physical-World Attacks on Machine Learning Models
- I. Evtimov, Kevin Eykholt, D. Song
- Computer ScienceArXiv
- 27 July 2017
This paper proposes a new attack algorithm--Robust Physical Perturbations (RP2)-- that generates perturbations by taking images under different conditions into account and can create spatially-constrained perturbation that mimic vandalism or art to reduce the likelihood of detection by a casual observer.
...
...