Learn More
In this paper, we present an improved method of anchor models for speaker verification. Anchor model is the method that represent a speaker by his relativity of a set of other speakers, called anchor speakers. It was firstly introduced for speaker indexing in large audio database. We suggest a rank based metric for the measurement of speaker character(More)
—Speech with various emotions aggravates the performance of speaker recognition system. The existing speaker modeling disregards the match of the emotional state between training and testing speech, and the systems suffer the lapsus of the emotion recognition as to practical application. We propose an alternative approach that exploits the prosodic(More)
In this paper, we attempt to use pitch contour in Mandarin isolated words recognition by matching candidate pitch contour with a set of templates. To apply it in speaker-independent speech recognition, we propose a remedy algorithm to weaken speaker information in pitch contour. Speaker-dependent and - independent speech recognition experiments are carried(More)
BACKGROUND Propofol has been reported to have an inhibitory effect on ischemia/reperfusion (I/R) injury in various experimental models by reducing oxidative stress, protecting mitochondrial function and suppressing apoptosis. The aim of this study was to investigate the effect and mechanism of propofol on myocardial I/R injury in type 2 diabetic rats. (More)
In this paper, we present a handheld device called GeeAir for remotely controlling home appliances via a mixed modality of speech, gesture, joystick, button, and light. This solution is superior to the existing universal remote controllers in that it can be used by the users with physical and vision impairments in a natural manner. By combining diverse(More)
In this paper, a large emotional speech database MASC (Mandarin Affective Speech Corpus) is introduced. The database contains recordings of 68 native speakers (23 female and 45 male) and five kinds of emotional states: neutral, anger, elation, panic and sadness. Each speaker pronounces 5 phrases, 10 sentences for three times for each emotional states and 2(More)