DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
@inproceedings{Rajput2019DETOXAR, title={DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation}, author={Shashank Rajput and H. Wang and Zachary Charles and Dimitris Papailiopoulos}, booktitle={NeurIPS}, year={2019} }
To improve the resilience of distributed training to worst-case, or Byzantine node failures, several recent approaches have replaced gradient averaging with robust aggregation methods. Such techniques can have high computational costs, often quadratic in the number of compute nodes, and only have limited robustness guarantees. Other methods have instead used redundancy to guarantee robustness, but can only tolerate limited number of Byzantine failures. In this work, we present DETOX, a… CONTINUE READING
Supplemental Code
Github Repo
Via Papers with Code
DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
Figures, Tables, and Topics from this paper
28 Citations
Fast and Robust Distributed Learning in High Dimension
- Computer Science
- 2020 International Symposium on Reliable Distributed Systems (SRDS)
- 2020
ByzShield: An Efficient and Robust System for Distributed Training
- Computer Science, Mathematics
- ArXiv
- 2020
- 1
- Highly Influenced
- PDF
Better scalability under potentially heavy-tailed gradients
- Computer Science, Mathematics
- ArXiv
- 2020
- 1
- PDF
Randomized Reactive Redundancy for Byzantine Fault-Tolerance in Parallelized Learning
- Computer Science
- ArXiv
- 2019
- 1
- Highly Influenced
- PDF
Resilient Distributed Diffusion for Multi-Robot Systems Using Centerpoint
- Computer Science
- RSS 2020
- 2020
- 3
- PDF
Federated Variance-Reduced Stochastic Gradient Descent With Robustness to Byzantine Attacks
- Computer Science, Mathematics
- IEEE Transactions on Signal Processing
- 2020
- 10
- PDF
References
SHOWING 1-10 OF 28 REFERENCES
DRACO: Byzantine-resilient Distributed Training via Redundant Gradients
- Mathematics, Computer Science
- ICML
- 2018
- 91
- PDF
AGGREGATHOR: Byzantine Machine Learning via Robust Gradient Aggregation
- Computer Science
- MLSys
- 2019
- 30
- PDF
Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent
- Computer Science
- NIPS
- 2017
- 224
- Highly Influential
- PDF
Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates
- Computer Science, Mathematics
- ICML
- 2018
- 185
- Highly Influential
- PDF
Data Encoding for Byzantine-Resilient Distributed Gradient Descent
- Computer Science
- 2018 56th Annual Allerton Conference on Communication, Control, and Computing (Allerton)
- 2018
- 10
Defending Against Saddle Point Attack in Byzantine-Robust Distributed Learning
- Computer Science, Mathematics
- ICML
- 2019
- 34
- PDF
Zeno: Distributed Stochastic Gradient Descent with Suspicion-based Fault-tolerance
- Computer Science
- ICML
- 2019
- 39
- PDF