Author pages are created from data sourced from our academic publisher partnerships and public sources.
Share This Author
Can you 'read' tongue movements? Evaluation of the contribution of tongue display to speech understanding
- P. Badin, Y. Tarabalka, F. Elisei, G. Bailly
- Psychology, Computer ScienceSpeech Commun.
- 1 December 2007
A set of audiovisual VCV stimuli with an audiovISual talking head that can display all speech articulators, including tongue, in an augmented speech mode and conclude that these tongue reading capabilities could be used for applications in the domains of speech therapy for speech retarded children, of perception and production rehabilitation of hearing impaired children, and of pronunciation training for second language learners.
SFC: A trainable prosodic model
This paper describes the original training procedure that enables the model-constrained and data-driven system to generate prosody from metalinguistic information to consider the prosodic continuum as the superposition of multiple elementary overlapping multiparametric contours.
Three-dimensional linear articulatory modeling of tongue, lips and face, based on MRI and video images
- P. Badin, G. Bailly, L. Revéret, M. Baciu, C. Segebarth, C. Savariaux
- Computer ScienceJ. Phonetics
- 1 July 2002
The geometry of these vocal organs is measured on one subject uttering a corpus of sustained articulations in French to imply that most 3D features such as tongue groove or lateral channels can be controlled by articulatory parameters defined for the midsagittal model.
I Reach Faster When I See You Look: Gaze Effects in Human–Human and Human–Robot Face-to-Face Cooperation
- Jean-David Boucher, U. Pattacini, +5 authors J. Ventre-Dominey
- Computer Science, MedicineFront. Neurorobot.
- 31 January 2012
The current research provides the robot with the capability to produce appropriate speech and gaze cues in the context of human-robot cooperation tasks, and demonstrates the pertinence of these cues in terms of statistical measures of action times for humans in thecontext of a cooperative task, as gaze significantly facilitates cooperation.
An Audiovisual Talking Head for Augmented Speech Generation: Models and Animations Based on a Real Speaker's Articulatory Data
A methodology developed to derive three-dimensional models of speech articulators from volume MRI and multiple view video images acquired on one speaker with promising applications in the domain of speech therapy for speech retarded children, perception and production rehabilitation of hearing impaired children, and pronunciation training for second language learners.
Linear degrees of freedom in speech production: analysis of cineradio- and labio-film data and articulatory-acoustic modeling.
- D. Beautemps, P. Badin, G. Bailly
- Mathematics, MedicineThe Journal of the Acoustical Society of America
- 5 June 2001
The following contribution addresses several issues concerning speech degrees of freedom in French oral vowels, stop, and fricative consonants based on an analysis of tongue and lip shapes extracted…
Learning to speak. Sensori-motor control of speech movements
- G. Bailly
- Philosophy, Computer ScienceSpeech Commun.
- 1 August 1997
Abstract This paper shows how an articulatory model, able to produce acoustic signals from articulatory motion, can learn to speak i.e. coordinate its movements in such a way that it utters…
LIPS2008: visual speech synthesis challenge
An overview of LIPS2008: Visual Speech Synthesis Challenge is presented, before describing the capture and preparation of the training data, the evaluation framework, and conclude with an outline of possible directions for standardising the evaluation of talking heads.
A three-dimensional linear articulatory model based on MRI data
A 3D statistical articulatory model has been built using guided Principal Component Analysis based on a set of 3D vocal tract images obtained by MRI, and a first evaluation of the model in terms of area functions and formants is presented.
Close Shadowing Natural Versus Synthetic Speech
- G. Bailly
- Computer ScienceInt. J. Speech Technol.
Preliminary results show that speakers are able to follow natural stimuli with an average delay of 70 ms whereas this delay typically exceeds 100 ms for stimuli produced by text-to-speech systems.