Audio-visual speech recognition using deep learning

@article{Noda2014AudiovisualSR,
  title={Audio-visual speech recognition using deep learning},
  author={Kuniaki Noda and Yuki Yamaguchi and Kazuhiro Nakadai and Hiroshi G. Okuno and Tetsuya Ogata},
  journal={Applied Intelligence},
  year={2014},
  volume={42},
  pages={722-737}
}
Audio-visual speech recognition (AVSR) system is thought to be one of the most promising solutions for reliable speech recognition, particularly when the audio is corrupted by noise. However, cautious selection of sensory features is crucial for attaining high recognition performance. In the machine-learning community, deep learning approaches have recently attracted increasing attention because deep neural networks can effectively extract robust latent features that enable various recognition… CONTINUE READING
Highly Cited
This paper has 138 citations. REVIEW CITATIONS
Recent Discussions
This paper has been referenced on Twitter 24 times over the past 90 days. VIEW TWEETS

Citations

Publications citing this paper.
Showing 1-10 of 73 extracted citations

138 Citations

02040602015201620172018
Citations per Year
Semantic Scholar estimates that this publication has 138 citations based on the available data.

See our FAQ for additional information.

References

Publications referenced by this paper.
Showing 1-10 of 50 references

Similar Papers

Loading similar papers…