mixup: Beyond Empirical Risk Minimization
- Hongyi Zhang, Moustapha Cissé, Y. Dauphin, David Lopez-Paz
- Computer ScienceInternational Conference on Learning…
- 25 October 2017
This work proposes mixup, a simple learning principle that trains a neural network on convex combinations of pairs of examples and their labels, which improves the generalization of state-of-the-art neural network architectures.
Countering Adversarial Images using Input Transformations
- Chuan Guo, Mayank Rana, Moustapha Cissé, L. V. D. Maaten
- Computer ScienceInternational Conference on Learning…
- 2018
This paper investigates strategies that defend against adversarial-example attacks on image-classification systems by transforming the inputs before feeding them to the system, and shows that total variance minimization and image quilting are very effective defenses in practice, when the network is trained on transformed images.
Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring
- Yossi Adi, Carsten Baum, Moustapha Cissé, Benny Pinkas, Joseph Keshet
- Computer ScienceUSENIX Security Symposium
- 13 February 2018
This work presents an approach for watermarking Deep Neural Networks in a black-box way, and shows experimentally that such a watermark has no noticeable impact on the primary task that the model is designed for.
Parseval Networks: Improving Robustness to Adversarial Examples
- Moustapha Cissé, Piotr Bojanowski, Edouard Grave, Y. Dauphin, Nicolas Usunier
- Computer ScienceInternational Conference on Machine Learning
- 28 April 2017
It is shown that Parseval networks match the state-of-the-art in terms of accuracy on CIFAR-10/100 and Street View House Numbers while being more robust than their vanilla counterpart against adversarial examples.
Efficient softmax approximation for GPUs
- Edouard Grave, Armand Joulin, Moustapha Cissé, David Grangier, H. Jégou
- Computer ScienceInternational Conference on Machine Learning
- 14 September 2016
This work proposes an approximate strategy to efficiently train neural network based language models over very large vocabularies by exploiting the unbalanced word distribution to form clusters that explicitly minimize the expectation of computational complexity.
Houdini: Fooling Deep Structured Prediction Models
- Moustapha Cissé, Yossi Adi, N. Neverova, Joseph Keshet
- Computer ScienceArXiv
- 17 July 2017
This work introduces a novel flexible approach named Houdini for generating adversarial examples specifically tailored for the final performance measure of the task considered, be it combinatorial and non-decomposable.
Fooling End-To-End Speaker Verification With Adversarial Examples
- Felix Kreuk, Yossi Adi, Moustapha Cissé, Joseph Keshet
- Computer ScienceIEEE International Conference on Acoustics…
- 10 January 2018
This paper presents white-box attacks on a deep end-to-end network that was either trained on YOHO or NTIMIT, and shows that one can significantly decrease the accuracy of a target system even when the adversarial examples are generated with different system potentially using different features.
Unbounded cache model for online language modeling with open vocabulary
- Edouard Grave, Moustapha Cissé, Armand Joulin
- Computer ScienceNIPS
- 7 November 2017
This paper uses a large scale non-parametric memory component that stores all the hidden activations seen in the past and leverages recent advances in approximate nearest neighbor search and quantization algorithms to store millions of representations while searching them efficiently.
ConvNets and ImageNet Beyond Accuracy: Explanations, Bias Detection, Adversarial Examples and Model Criticism
- Pierre Stock, Moustapha Cissé
- Computer ScienceArXiv
- 30 November 2017
It is shown that explanations can mitigate the impact of misclassified adversarial examples from the perspective of the end-user and a novel tool for uncovering the undesirable biases learned by a model is introduced.
ADIOS: Architectures Deep In Output Space
- Moustapha Cissé, Maruan Al-Shedivat, Samy Bengio
- Computer ScienceInternational Conference on Machine Learning
- 19 June 2016
This paper proposes to make use of the underlying structure of binary classification by learning to partition the labels into a Markov Blanket Chain and then applying a novel deep architecture that exploits the partition.
...
...