Juan Manuel Montero-Martínez

Learn More
Modern Speech synthesisers have achieved a high degree of intelligibility, but can not be regarded as natural-sounding devices. In order to decrease the monotony of synthetic speech, the implementation of emotional effects is now being progressively considered. This paper presents a through study of emotional speech in Spanish, and its application to TTS,(More)
This article describes an architecture for translating speech into Spanish Sign Language (SSL). The architecture proposed is made up of four modules: speech recognizer, semantic analysis, gesture sequence generation and gesture playing. For the speech recognizer and the semantic analysis modules, we use software developed by IBM and CSLR (Center for Spoken(More)
+P VJKU RCRGT YG FGUETKDG VJG UVGRU ECTTKGF QWV HQT FGXGNQRKPI C 4CKNYC[ +PHQTOCVKQP 5GTXKEG HQT 5RCPKUJ 6JKU YQTM KPVTQFWEGU C OGVJQFQNQI[ HQT FGUKIPKPI FKCNQIWG OCPCIGTU KP URQMGP FKCNQIWG U[UVGOU HQT TGUVTKEVGF FQOCKPU +P VJKU OGVJQFQNQI[ UGXGTCN UQWTEGU QH KPHQTOCVKQP CTG EQODKPGF KPVWKVKQP QDUGTXCVKQP CPF UKOWNCVKQP HQT FGHKPKPI UGXGTCN FKCNQIWG(More)
This paper describes the development of and the first experiments in a Spanish to sign language translation system in a real domain. The developed system focuses on the sentences spoken by an official when assisting people applying for, or renewing their Identity Card. The system translates official explanations into Spanish Sign Language (LSE: Lengua de(More)
+P VJKU RCRGT YG RTQRQUG CP CRRTQCEJ HQT FGUKIPKPI VJG EQPHKTOCVKQP UVTCVGIKGU KP C 4CKNYC[ +PHQTOCVKQP U[UVGO HQT 5RCPKUJ DCUGF QP EQPHKFGPEG OGCUWTGU QDVCKPGF HTQO TGEQIPKVKQP 9G CNUQ RTGUGPV UGXGTCN GTTQT TGEQXGT CPF WUGT OQFGNNKPI VGEJPKSWGU KPEQTRQTCVGF KP VJKU U[UVGO +P VJG HKGNF GXCNWCVKQP KV KU UJQYP VJCV OQTG VJCP QH VJG EQPHKTOCVKQPU YGTG KORNKEKV(More)
We have applied two state-of-the-art speech synthesis techniques (unit selection and HMM-based synthesis) to the synthesis of emotional speech. A series of carefully designed perceptual tests to evaluate speech quality, emotion identification rates and emotional strength were used for the six emotions which we recorded – happiness, sadness, anger, surprise,(More)
The growing interest in emotional speech synthesis urges effective emotion conversion techniques to be explored. This paper estimates the relevance of three speech components (spectral envelope, residual excitation and prosody) for synthesizing identifiable emotional speech, in order to be able to customize the voice conversion techniques to the specific(More)
This paper describes the first experiments of a speech to sign language translation system in a real domain. The developed system is focused on the sentences spoken by an officer when assisting people in applying for, or renewing the National Identification Document (NID) and the Passport. This system translates officer explanations into sign language for(More)
In this paper, we present the work done in language identification for two air traffic control speech recognizers, one for continuous speech and the other one for a command interface. The system is able to distinguish between Spanish and English. We will confirm the advantage of using PPRLM over PRLM. All previous studies show that PPRLM is the technique(More)
In this paper we describe the acoustic emotion recognition system built at the Speech Technology Group of the Universidad Politecnica de Madrid (Spain) to participate in the INTERSPEECH 2009 Emotion Challenge. Our proposal is based on the use of a Dynamic Bayesian Network (DBN) to deal with the temporal modelling of the emotional speech information. The(More)