Learn More
This paper presents a new approach for designing a concatenative text-to-speech (TTS) system based on multi-domain unit selection. The method achieves good synthetic quality with reasonable computational cost for a general-purpose TTS system. The architecture of the multi-domain database and the text classification algorithm for domain assignment are the(More)
This paper describes the methodology used for validating the results obtained in a study about acoustical modelling of emotional expression in Castilian Spanish. We have obtained a set of rules that describes the behaviour of the most significant parameters of speech related with the emotional expression. The validation of the results of the study has been(More)
Hidden Markov Models based text-to-speech (HMM-TTS) synthesis is a technique for generating speech from trained statistical models where spectrum, pitch and durations of basic speech units are modelled altogether. The aim of this work is to describe a Spanish HMM-TTS system using CBR as a F0 esti-mator, analysing its performance objectively and(More)
In this work, the capability of voice quality parameters to discriminate among different expressive speech styles is analyzed. To that effect, the data distribution of these parameters, directly measured from the acoustic speech signal, is used to train a Linear Discriminant Analysis that conducts an automatic classification. As a result, the most relevant(More)
An artificial vision system for vehicles is proposed in this article to alert drivers of potential head on collisions. It is capable of detecting any type of frontal collision from any type of obstacle that may present itself in a vehiclepsilas path. The system operates based on a sequence of algorithms whose images are recorded on a camera located in the(More)
 This paper proposes an approach to transform speech from a neutral style into other expressive styles using both prosody and voice quality (VoQ). The main aim is to validate the usefulness of VoQ in the enhancement of expressive synthetic speech. A Harmonic plus Noise Model (HNM) is used to modify speech following a set of rules extracted from an(More)
Detection of affective states in speech could improve the way users interact with electronic devices. However the analysis of speech at the acoustic level could be not enough to determine the emotion of a user speaking in a realistic scenario. In this paper we analysed the spontaneous speech recordings of the FAU Aibo Corpus at the acoustic and linguistic(More)
Induced emotions (chairman A. Batliner) 9.05 The Sensitive Artificial Listener: an induction technique for generating emotionally coloured conversation. Acted versus spontaneous emotions (chairman R. Cowie) 10.45 Anger detection performances based on prosodic and acoustic cues in several corpora Laurence Vidrascu, Laurence Devillers, LIMSI-CNRS, France(More)
This paper presents the validation of the expres-siveness of an acted corpus produced to be used in speech synthesis, as this kind of emotional speech can be rather lacking in authenticity. The goal is to obtain a system which is able to prune bad utterances from an expressiveness point of view. The results from a previous subjective test are used for the(More)
This paper presents the perceptual experiments that were carried out in order to validate the methodology of transforming expressive speech styles using voice quality (VoQ) parameters modelling, along with the well-known prosody (F 0, duration, and energy), from a neutral style into a number of expressive ones. The main goal was to validate the usefulness(More)