Soft Threshold Weight Reparameterization for Learnable Sparsity
@article{Kusupati2020SoftTW, title={Soft Threshold Weight Reparameterization for Learnable Sparsity}, author={Aditya Kusupati and V. Ramanujan and Raghav Somani and Mitchell Wortsman and Prateek Jain and S. Kakade and A. Farhadi}, journal={ArXiv}, year={2020}, volume={abs/2002.03231} }
Sparsity in Deep Neural Networks (DNNs) is studied extensively with the focus of maximizing prediction accuracy given an overall parameter budget. Existing methods rely on uniform or heuristic non-uniform sparsity budgets which have sub-optimal layer-wise parameter allocation resulting in a) lower prediction accuracy or b) higher inference cost (FLOPs). This work proposes Soft Threshold Reparameterization (STR), a novel use of the soft-threshold operator on DNN weights. STR smoothly induces… Expand
Figures, Tables, and Topics from this paper
25 Citations
Learning N: M Fine-grained Structured Sparse Neural Networks From Scratch
- Computer Science
- ArXiv
- 2021
- 1
- Highly Influenced
- PDF
Hierarchical Adaptive Lasso: Learning Sparse Neural Networks with Shrinkage via Single Stage Training
- Computer Science
- ArXiv
- 2020
Do We Actually Need Dense Over-Parameterization? In-Time Over-Parameterization in Sparse Training
- Computer Science
- ArXiv
- 2021
- 1
- Highly Influenced
- PDF
References
SHOWING 1-10 OF 65 REFERENCES
Parameter Efficient Training of Deep Convolutional Neural Networks by Dynamic Sparse Reparameterization
- Computer Science, Mathematics
- ICML
- 2019
- 65
- Highly Influential
- PDF
The State of Sparsity in Deep Neural Networks
- Computer Science, Mathematics
- ArXiv
- 2019
- 181
- Highly Influential
- PDF
To prune, or not to prune: exploring the efficacy of pruning for model compression
- Computer Science, Mathematics
- ICLR
- 2018
- 332
- Highly Influential
- PDF
Sparse Convolutional Neural Networks
- Computer Science
- 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
- 2015
- 493
- PDF