Deep Learning Acceleration with Neuron-to-Memory Transformation
@article{Imani2020DeepLA, title={Deep Learning Acceleration with Neuron-to-Memory Transformation}, author={M. Imani and Mohammad Samragh Razlighi and Y. Kim and Saransh Gupta and F. Koushanfar and T. Simunic}, journal={2020 IEEE International Symposium on High Performance Computer Architecture (HPCA)}, year={2020}, pages={1-14} }
Deep neural networks (DNN) have demonstrated effectiveness for various applications such as image processing, video segmentation, and speech recognition. Running state-of-theart DNNs on current systems mostly relies on either generalpurpose processors, ASIC designs, or FPGA accelerators, all of which suffer from data movements due to the limited on-chip memory and data transfer bandwidth. In this work, we propose a novel framework, called RAPIDNN, which performs neuron-to-memory transformation… Expand
Figures, Tables, and Topics from this paper
7 Citations
DUAL: Acceleration of Clustering Algorithms using Digital-based Processing In-Memory
- Computer Science
- 2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)
- 2020
- 3
- PDF
SpAtten: Efficient Sparse Attention Architecture with Cascade Token and Head Pruning
- Computer Science
- ArXiv
- 2020
- 1
- PDF
Hardware Acceleration of Sparse and Irregular Tensor Computations of ML Models: A Survey and Insights
- Computer Science
- ArXiv
- 2020
- 3
- PDF
Cross-layer Design for Computing-in-Memory: From Devices, Circuits, to Architectures and Applications
- Computer Science
- 2021 26th Asia and South Pacific Design Automation Conference (ASP-DAC)
- 2021
Implementing binary neural networks in memory with approximate accumulation
- Computer Science
- ISLPED
- 2020
- PDF
References
SHOWING 1-10 OF 92 REFERENCES
NNPIM: A Processing In-Memory Architecture for Neural Network Acceleration
- Computer Science
- IEEE Transactions on Computers
- 2019
- 14
- PDF
ISAAC: A Convolutional Neural Network Accelerator with In-Situ Analog Arithmetic in Crossbars
- Computer Science
- 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA)
- 2016
- 746
- Highly Influential
- PDF
From high-level deep neural models to FPGAs
- Computer Science
- 2016 49th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO)
- 2016
- 256
- PDF
PipeLayer: A Pipelined ReRAM-Based Accelerator for Deep Learning
- Computer Science
- 2017 IEEE International Symposium on High Performance Computer Architecture (HPCA)
- 2017
- 296
- Highly Influential
- PDF
DaDianNao: A Machine-Learning Supercomputer
- Computer Science
- 2014 47th Annual IEEE/ACM International Symposium on Microarchitecture
- 2014
- 894
- Highly Influential
- PDF
Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks
- Computer Science
- FPGA
- 2015
- 1,173
- PDF
Optimizing Loop Operation and Dataflow in FPGA Acceleration of Deep Convolutional Neural Networks
- Computer Science
- FPGA
- 2017
- 193
EIE: Efficient Inference Engine on Compressed Deep Neural Network
- Computer Science
- 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA)
- 2016
- 1,439
- PDF
Neural Cache: Bit-Serial In-Cache Acceleration of Deep Neural Networks
- Computer Science
- 2018 ACM/IEEE 45th Annual International Symposium on Computer Architecture (ISCA)
- 2018
- 116
- PDF