TinyTL: Reduce Memory, Not Parameters for Efficient On-Device Learning
@inproceedings{Cai2020TinyTLRM, title={TinyTL: Reduce Memory, Not Parameters for Efficient On-Device Learning}, author={H. Cai and Chuang Gan and Ligeng Zhu and Song Han}, booktitle={NeurIPS}, year={2020} }
On-device learning enables edge devices to continually adapt the AI models to new data, which requires a small memory footprint to fit the tight memory constraint of edge devices. Existing work solves this problem by reducing the number of trainable parameters. However, this doesn't directly translate to memory saving since the major bottleneck is the activations, not parameters. In this work, we present Tiny-Transfer-Learning (TinyTL) for memory-efficient on-device learning. TinyTL freezes the… CONTINUE READING
Supplemental Presentations
Presentation Slides
Figures and Tables from this paper
References
SHOWING 1-10 OF 60 REFERENCES
MemNet: Memory-Efficiency Guided Neural Architecture Search with Augment-Trim learning
- Mathematics, Computer Science
- ArXiv
- 2019
- 3
- PDF
ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware
- Computer Science, Mathematics
- ICLR
- 2019
- 638
- PDF
MnasNet: Platform-Aware Neural Architecture Search for Mobile
- Computer Science
- 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
- 2019
- 830
- Highly Influential
- PDF
K For The Price Of 1: Parameter Efficient Multi-task And Transfer Learning
- Computer Science, Mathematics
- ICLR
- 2019
- 19
- PDF
Dynamic Sparse Graph for Efficient Deep Learning
- Computer Science, Mathematics
- ICLR
- 2019
- 17
- Highly Influential
- PDF
Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding
- Computer Science
- ICLR
- 2016
- 4,207
- Highly Influential
- PDF