Corpus ID: 12670695

MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications

@article{Howard2017MobileNetsEC,
  title={MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications},
  author={Andrew G. Howard and Menglong Zhu and Bo Chen and Dmitry Kalenichenko and Weijun Wang and Tobias Weyand and Marco Andreetto and Hartwig Adam},
  journal={ArXiv},
  year={2017},
  volume={abs/1704.04861}
}
We present a class of efficient models called MobileNets for mobile and embedded vision applications. [...] Key Method These hyper-parameters allow the model builder to choose the right sized model for their application based on the constraints of the problem. We present extensive experiments on resource and accuracy tradeoffs and show strong performance compared to other popular models on ImageNet classification. We then demonstrate the effectiveness of MobileNets across a wide range of applications and use…Expand
Thin MobileNet: An Enhanced MobileNet Architecture
  • Debjyoti Sinha, M. El-Sharkawy
  • Computer Science
  • 2019 IEEE 10th Annual Ubiquitous Computing, Electronics & Mobile Communication Conference (UEMCON)
  • 2019
TLDR
Three hybrid MobileNet architectures are proposed which has improved accuracy along-with reduced size, lesser number of layers, lower average computation time and very less overfitting as compared to the baseline MobileNet v1. Expand
MobileSR: Efficient Convolutional Neural Network for Super-resolution
TLDR
This paper proposes a parallel-group convolution, which uses 25% computation of the standard convolutions, and develops an efficient light-weight convolutional neural network named MobileSR for super-resolution. Expand
ANTNets: Mobile Convolutional Neural Networks for Resource Efficient Image Classification
TLDR
Under budget constraints such as computational cost (MAdds) and the parameter count, a novel basic architectural block, ANTBlock, is proposed, which boosts the representational power by modeling, in a high dimensional space, interdependency of channels between a depthwise convolution layer and a projection layer in the ANTBlocks. Expand
PydMobileNet: Improved Version of MobileNets with Pyramid Depthwise Separable Convolution
TLDR
An improved version of MobileNet, called Pyramid Mobile Network, which is more flexible in fine-tuning the trade-off between accuracy, latency and model size than MobileNets and evaluated on two highly competitive object recognition benchmark datasets. Expand
Ultra-thin MobileNet
TLDR
This paper proposes some modifications to the existing baseline MobileNet architecture to make it more efficient and suitable to be deployed on real-time embedded platforms and names the new DNN architecture as Ultra-thin MobileNet having a size of 3.9 MB only which is deployable in real- time embedded processors with limited memory and power. Expand
Quantization Friendly MobileNet (QF-MobileNet) Architecture for Vision Based Applications on Embedded Platforms
TLDR
This paper proposes Quantization Friendly MobileNet (QF-MobileNet) architecture, an optimized DNN model for memory and accuracy for vision-based applications on embedded platforms and intends to explore quantization aware training of DNN algorithms to achieve optimized resource utilization and inference accuracy. Expand
MoBiNet: A Mobile Binary Network for Image Classification
TLDR
A novel neural network architecture, namely MoBi- Net - Mobile Binary Network in which skip connections are manipulated to prevent information loss and vanishing gradient, thus facilitate the training process and results in an effectively small model while keeping the accuracy comparable to existing ones. Expand
ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices
TLDR
An extremely computation-efficient CNN architecture named ShuffleNet is introduced, which is designed specially for mobile devices with very limited computing power (e.g., 10-150 MFLOPs), to greatly reduce computation cost while maintaining accuracy. Expand
Slim MobileNet: An Enhanced Deep Convolutional Neural Network
TLDR
An enhanced version of MobileNet is proposed that verify the condition to be implemented on an embedded board while improved the accuracy and has fewer number of layers, improved accuracy while depreciating the overall size and lower average time compared to the MobileNet-V1 model. Expand
PydMobileNet: Pyramid Depthwise Separable Convolution Networks for Image Classification
  • Van-Thanh Hoang, K. Jo
  • Computer Science
  • 2019 IEEE 28th International Symposium on Industrial Electronics (ISIE)
  • 2019
TLDR
An improved version of MobileNet, called Pyramid Mobile Network, which is more flexible in fine-tuning the trade-off between accuracy, latency and model size than MobileNets and evaluated on two highly competitive object recognition benchmark datasets. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 39 REFERENCES
Quantized Convolutional Neural Networks for Mobile Devices
TLDR
This paper proposes an efficient framework, namely Quantized CNN, to simultaneously speed-up the computation and reduce the storage and memory overhead of CNN models. Expand
Very Deep Convolutional Networks for Large-Scale Image Recognition
TLDR
This work investigates the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting using an architecture with very small convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers. Expand
Rethinking the Inception Architecture for Computer Vision
TLDR
This work is exploring ways to scale up networks in ways that aim at utilizing the added computation as efficiently as possible by suitably factorized convolutions and aggressive regularization. Expand
Speeding up Convolutional Neural Networks with Low Rank Expansions
TLDR
Two simple schemes for drastically speeding up convolutional neural networks are presented, achieved by exploiting cross-channel or filter redundancy to construct a low rank basis of filters that are rank-1 in the spatial domain. Expand
Structured Transforms for Small-Footprint Deep Learning
TLDR
In keyword spotting applications in mobile speech recognition, the proposed unified framework to learn a broad family of structured parameter matrices that are characterized by the notion of low displacement rank are much more effective than standard linear low-rank bottleneck layers and nearly retain the performance of state of the art models. Expand
Caffe: Convolutional Architecture for Fast Feature Embedding
TLDR
Caffe provides multimedia scientists and practitioners with a clean and modifiable framework for state-of-the-art deep learning algorithms and a collection of reference models for training and deploying general-purpose convolutional neural networks and other deep models efficiently on commodity architectures. Expand
Speed/Accuracy Trade-Offs for Modern Convolutional Object Detectors
TLDR
A unified implementation of the Faster R-CNN, R-FCN and SSD systems is presented and the speed/accuracy trade-off curve created by using alternative feature extractors and varying other critical parameters such as image size within each of these meta-architectures is traced out. Expand
Going deeper with convolutions
We propose a deep convolutional neural network architecture codenamed Inception that achieves the new state of the art for classification and detection in the ImageNet Large-Scale Visual RecognitionExpand
Factorized Convolutional Neural Networks
TLDR
The proposed convolutional layer is composed of a low-cost single intra-channel convolution and a linear channel projection that can effectively preserve the spatial information and maintain the accuracy with significantly less computation. Expand
XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks
TLDR
The Binary-Weight-Network version of AlexNet is compared with recent network binarization methods, BinaryConnect and BinaryNets, and outperform these methods by large margins on ImageNet, more than \(16\,\%\) in top-1 accuracy. Expand
...
1
2
3
4
...