Trainable videorealistic speech animation

@article{Ezzat2002TrainableVS,
  title={Trainable videorealistic speech animation},
  author={Tony Ezzat and Gadi Geiger and Tomaso A. Poggio},
  journal={Sixth IEEE International Conference on Automatic Face and Gesture Recognition, 2004. Proceedings.},
  year={2002},
  pages={57-64}
}
We describe how to create with machine learning techniques a generative, speech animation module. A human subject is first recorded using a videocamera as he/she utters a predetermined speech corpus. After processing the corpus automatically, a visual speech module is learned from the data that is capable of synthesizing the human subject's mouth uttering entirely novel utterances that were not recorded in the original video. The synthesized utterance is re-composited onto a background sequence… CONTINUE READING

Citations

Publications citing this paper.
SHOWING 1-10 OF 29 CITATIONS

A Statistical Quality Model for Data-Driven Speech Animation

  • IEEE Transactions on Visualization and Computer Graphics
  • 2012
VIEW 11 EXCERPTS
CITES BACKGROUND & METHODS
HIGHLY INFLUENCED

Animating Lip-Sync Characters With Dominated Animeme Models

  • IEEE Transactions on Circuits and Systems for Video Technology
  • 2012
VIEW 9 EXCERPTS
CITES BACKGROUND & METHODS
HIGHLY INFLUENCED

Language Independent Speech Driven Facial Animation

  • 2008 IEEE Region 10 and the Third international Conference on Industrial and Information Systems
  • 2008
VIEW 3 EXCERPTS
CITES BACKGROUND
HIGHLY INFLUENCED

Animating visible speech and facial expressions

  • The Visual Computer
  • 2003
VIEW 3 EXCERPTS
CITES METHODS & BACKGROUND
HIGHLY INFLUENCED

Expressive talking avatar synthesis and animation

  • Multimedia Tools and Applications
  • 2015
VIEW 1 EXCERPT
CITES BACKGROUND

Realtime speech-driven facial animation using Gaussian Mixture Models

  • 2014 IEEE International Conference on Multimedia and Expo Workshops (ICMEW)
  • 2014
VIEW 1 EXCERPT
CITES BACKGROUND

Synthesizing real-time speech-driven facial animation

  • 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
  • 2014
VIEW 1 EXCERPT
CITES METHODS

References

Publications referenced by this paper.
SHOWING 1-10 OF 12 REFERENCES

Neural Networks for Pattern Recognition

C. M. Bishop
  • Clarendon Press, Oxford
  • 1995
VIEW 7 EXCERPTS
HIGHLY INFLUENTIAL

Trainable videorealistic facial animation

VIEW 5 EXCERPTS
HIGHLY INFLUENTIAL

EM Algorithms for PCA and SPCA

VIEW 6 EXCERPTS
HIGHLY INFLUENTIAL

Priors, stabilizers, and basis functions: From regularization to radial, tensor, and additive splines

F. Girosi, M. Jones, T. Poggio
  • Technical Report 1430,
  • 1993
VIEW 4 EXCERPTS
HIGHLY INFLUENTIAL

and R

X. Huang, F. Alleva, H.-W. Hon, M.-Y. Hwang, K.-F. Lee
  • Rosenfeld. The SPHINX- II speech recognition system: an overview (http://sourceforge.net/projects/cmusphinx/). Computer Speech and Language, 7(2):137–148
  • 1993
VIEW 6 EXCERPTS
HIGHLY INFLUENTIAL

Determining Optical Flow

VIEW 4 EXCERPTS
HIGHLY INFLUENTIAL

A text-to-audiovisual-speech synthesizer for French

  • Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP '96
  • 1996
VIEW 4 EXCERPTS
HIGHLY INFLUENTIAL

Similar Papers

Loading similar papers…