Batch normalized recurrent neural networks
@article{Laurent2016BatchNR, title={Batch normalized recurrent neural networks}, author={C{\'e}sar Laurent and G. Pereyra and Philemon Brakel and Y. Zhang and Yoshua Bengio}, journal={2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)}, year={2016}, pages={2657-2661} }
Recurrent Neural Networks (RNNs) are powerful models for sequential data that have the potential to learn long-term dependencies. However, they are computationally expensive to train and difficult to parallelize. Recent work has shown that normalizing intermediate representations of neural networks can significantly improve convergence rates in feed-forward neural networks [1]. In particular, batch normalization, which uses mini-batch statistics to standardize features, was shown to… CONTINUE READING
153 Citations
Investigation on the Combination of Batch Normalization and Dropout in BLSTM-based Acoustic Modeling for ASR
- Computer Science
- INTERSPEECH
- 2018
- 4
- PDF
Layer-Normalized LSTM for Hybrid-Hmm and End-To-End ASR
- Computer Science
- ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
- 2020
- 3
- PDF
A comprehensive study of batch construction strategies for recurrent neural networks in MXNet
- Computer Science, Mathematics
- ArXiv
- 2017
- 9
- PDF
Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN
- Computer Science
- 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
- 2018
- 251
- PDF
References
SHOWING 1-10 OF 26 REFERENCES
Speech recognition with deep recurrent neural networks
- Computer Science
- 2013 IEEE International Conference on Acoustics, Speech and Signal Processing
- 2013
- 5,673
- PDF
Bidirectional recurrent neural networks
- Computer Science
- IEEE Trans. Signal Process.
- 1997
- 3,832
- Highly Influential
- PDF
Scaling recurrent neural network language models
- Computer Science
- 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
- 2015
- 57
- PDF
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
- Computer Science
- ICML
- 2015
- 21,339
- Highly Influential
- PDF
Hybrid speech recognition with Deep Bidirectional LSTM
- Computer Science
- 2013 IEEE Workshop on Automatic Speech Recognition and Understanding
- 2013
- 1,067
- PDF
Dropout: a simple way to prevent neural networks from overfitting
- Computer Science
- J. Mach. Learn. Res.
- 2014
- 21,306
- PDF