Dynamic Adaptive DNN Surgery for Inference Acceleration on the Edge

@article{Hu2019DynamicAD,
  title={Dynamic Adaptive DNN Surgery for Inference Acceleration on the Edge},
  author={Chuang Hu and Wei Shun Bao and Huijun Zhao and FengMing Liu},
  journal={IEEE INFOCOM 2019 - IEEE Conference on Computer Communications},
  year={2019},
  pages={1423-1431}
}
Recent advances in deep neural networks (DNNs) have substantially improved the accuracy and speed of a variety of intelligent applications. Nevertheless, one obstacle is that DNN inference imposes heavy computation burden to end devices, but offloading inference tasks to the cloud causes transmission of a large volume of data. Motivated by the fact that the data size of some intermediate DNN layers is significantly smaller than that of raw input data, we design the DNN surgery, which allows… CONTINUE READING

Similar Papers

Citations

Publications citing this paper.

References

Publications referenced by this paper.
SHOWING 1-10 OF 21 REFERENCES

YOLO9000: Better, Faster, Stronger

  • 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
  • 2016
VIEW 7 EXCERPTS
HIGHLY INFLUENTIAL

Deep Residual Learning for Image Recognition

  • 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
  • 2015
VIEW 6 EXCERPTS
HIGHLY INFLUENTIAL

Distributed Deep Neural Networks Over the Cloud, the Edge and End Devices

  • 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS)
  • 2017
VIEW 1 EXCERPT

DeepX: A Software Accelerator for Low-Power Deep Learning Inference on Mobile Devices

  • 2016 15th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN)
  • 2016
VIEW 1 EXCERPT