Problems of the Automatic Emotion Recognitions in Spontaneous Speech; An Example for the Recognition in a Dispatcher Center

@inproceedings{Vicsi2010ProblemsOT,
  title={Problems of the Automatic Emotion Recognitions in Spontaneous Speech; An Example for the Recognition in a Dispatcher Center},
  author={Kl{\'a}ra Vicsi and D{\'a}vid Sztah{\'o}},
  booktitle={COST 2102 Training School},
  year={2010}
}
Numerous difficulties, in the examination of emotions occurring in continuous spontaneous speech, are discussed in this paper, than different emotion recognition experiments are presented, using clauses as the recognition unit. In a testing experiment it was examined that what kind of acoustical features are the most important for the characterization of emotions, using spontaneous speech database. An SVM classifier was built for the classification of 4 most frequent emotions. It was found that… 

Speech activity detection and automatic prosodic processing unit segmentation for emotion recognition

In speech communication emotions play a great role in expressing information. These emotions are partly given as reactions to our environment, to our partners during a conversation. Understanding

Gender dependent word-level emotion detection using global spectral speech features

TLDR
Results showed that, using all 13 MFCC coefficients provide better classification results than other combinations of MFCC coefficient for the mentioned emotions.

Language Independent Detection Possibilities of Depression by Speech

In this study, acoustic-phonetic analysis of continuous speech and statistical analyses were performed in order to find parameters in depressed speech that show significant differences compared to a

Assessment of pathological speech prosody based on automatic stress detection and phrasing approaches

TLDR
The results show that healthy and pathological samples can be separated from each other by means of these prosodic analysers, and deep neural network or support vector machine based classifiers built on top of them.

Examination of the sensitivity of acoustic-phonetic parameters of speech to depression

Examination of the sensitivity of acoustic-phonetic parameters of speech to depression was started in the Laboratory of Speech Acoustics. In the presented research work some segmental parameters of

Seasonal Affective Disorder Speech Detection on the Base of Acoustic-Phonetic Speech Parameters

The development of an online monitoring system is shown in order to track the physiological and cognitive condition of crew members of the Concordia Research Station in Antarctica, with specific

Thinking about the present and future of the complex speech recognition

  • K. Vicsi
  • Computer Science
    2012 IEEE 3rd International Conference on Cognitive Infocommunications (CogInfoCom)
  • 2012
TLDR
The paper gives a short introduction of the principles of this speech recognition technology today and highlights the fact that these systems in the market are only speech-to-text transformers giving only a word chain at the output, where the speech prosody, speech emotion, speech style and more other information are not involved.

Physiological and Cognitive Status Monitoring on the Base of Acoustic-Phonetic Speech Parameters

In this paper the development of an online monitoring system is shown in order to track physiological and cognitive condition of crew members of the Concordia Research Station in Antarctica, with

Automatic identification of successful phone calls in call centers based on dialogue analysis

  • H. AtassiZ. Smekal
  • Computer Science
    2014 5th IEEE Conference on Cognitive Infocommunications (CogInfoCom)
  • 2014
TLDR
Novel approach to automatic identification of successful phone calls in call centers by means of dialogue features, extracted from four cues namely hesitation, reaction, Interruption and cumulative voice activity suggested to have a strong discriminative power.

References

SHOWING 1-10 OF 17 REFERENCES

Speech Emotion Perception by Human and Machine

TLDR
An automatic emotion recognition system was developed and recordings of 3 Hungarian sentences with 8 basic emotions pronounced by nonprofessional speakers showed similar recognition results as an earlier perception test obtained with professional actors/actresses.

A rule-based emotion-dependent feature extraction method for emotion analysis from speech.

TLDR
A rule-based method to determine emotion-dependent features, which are defined from high-level features derived from the statistical measurements of prosodic parameters of speech, was used to analyze emotional speech from five male and four female speakers in order to find any similarities and differences among individual speakers.

Emotional speech: Towards a new generation of databases

Study on Speaker-Independent Emotion Recognition from Speech on Real-World Data

TLDR
Experimental results demonstrate significant differences on recognizing emotions on acted/real-world speech.

Using prosody for the improvement of ASR - sentence modality recognition

TLDR
A semantic level processing, built on supra-segmental parameters, of types of modalities of sentences in Hungarian and German, using HMM models of modality types of sentences.

A database of German emotional speech

TLDR
A database of emotional speech that was evaluated in a perception test regarding the recognisability of emotions and their naturalness and can be accessed by the public via the internet.

An objective and subjective study of the role of semantics and prosodic features in building corpora for emotional TTS

TLDR
Experiments of automatic recognition of the emotions considered in this paper show that prosody is an important emotional indicator, but cannot be the only manipulated parameter in an emotional TTS system-at least not for all the emotions.

Getting to the Heart of the Matter; Speech is More than Just the Expression of Text or Language

This talk addresses the current needs for so-called emotion in speech, but points out that the issue is better described as the expression of relationships and attitudes rather than the currently

Individual Traits of Speaking Style and Speech Rhythm in a Spoken Discourse

TLDR
Several different styles of interaction can be automatically distinguished in the conversational speech data from their timing patterns, and noticeable differences were found to correlate well with degree of familiarity with the interlocutor.

The Perceptual and Cognitive Role of Visual and Auditory Channels in Conveying Emotional Information

TLDR
A comparative analysis of the subjective perception of emotional states by the visual and auditory channels considered either singularly or in combination, always in the non-verbal modality reveals that the audio and visual components of emotional messages convey much the same amount of information either separately or in combinations, suggesting that each channel performs a robust encoding of the emotional features.