Corpus ID: 8571120

Construction of Japanese Audio-Visual Emotion Database and Its Application in Emotion Recognition

@inproceedings{Lubis2016ConstructionOJ,
  title={Construction of Japanese Audio-Visual Emotion Database and Its Application in Emotion Recognition},
  author={Nurul Lubis and Randy Gomez and Sakriani Sakti and Keisuke Nakamura and Koichiro Yoshino and Satoshi Nakamura and Kazuhiro Nakadai},
  booktitle={LREC},
  year={2016}
}
Emotional aspects play a vital role in making human communication a rich and dynamic experience. As we introduce more automated system in our daily lives, it becomes increasingly important to incorporate emotion to provide as natural an interaction as possible. To achieve said incorporation, rich sets of labeled emotional data is prerequisite. However, in Japanese, existing emotion database is still limited to unimodal and bimodal corpora. Since emotion is not only expressed through speech, but… Expand

Figures, Tables, and Topics from this paper

Developing a Thai emotional speech corpus from Lakorn (EMOLA)
TLDR
The design of a Thai emotional speech corpus (namely EMOLA), its construction and annotation process, and its analysis are presented. Expand
Building Audio-Visual Phonetically Annotated Arabic Corpus for Expressive Text to Speech
TLDR
The corpus evaluation process showed that happy, anger and inquiry emotions were better recognized visually than audibly and the audio visual evaluation scores were higher than the audibly-based scores. Expand

References

SHOWING 1-10 OF 18 REFERENCES
The Vera am Mittag German audio-visual emotional speech database
TLDR
This contribution presents a recently collected database of spontaneous emotional speech in German which is being made available to the research community and provides emotion labels for a great part of the data. Expand
Emotion and Its Triggers in Human Spoken Dialogue: Recognition and Analysis
TLDR
A method of automatic recognition of emotion using support vector machine (SVM) and further analysis regarding emotion triggers is presented and the result shows performance that surpasses random guessing accuracy. Expand
The INTERSPEECH 2009 emotion challenge
TLDR
The challenge, the corpus, the features, and benchmark results of two popular approaches towards emotion recognition from speech, and the FAU Aibo Emotion Corpus are introduced. Expand
A corpus-based speech synthesis system with emotion
TLDR
The results show that the proposed method can synthesize speech with a high intelligibility and gives a favorable impression, and has developed a workable text-to-speech system with emotion to support the immediate needs of nonspeaking individuals. Expand
Emotion recognition in spontaneous emotional speech for anonymity-protected voice chat systems
TLDR
For the purpose of determining emotion recognition by acoustic information, natural dialogs made by two or three players of online games are recorded to construct an emotional speech database and discriminant analysis to discriminate one emotion from the others is conducted. Expand
A Corpus for the Study of Multimodal Emotional Expressions
This paper introduces the GEMEP (Geneva Multimodal Emotion Portrayals) corpus, a new repository of portrayed emotional expressions. Corpora of acted portrayals tend to include portrayals of veryExpand
Survey on speech emotion recognition: Features, classification schemes, and databases
TLDR
A survey of speech emotion classification addressing three important aspects of the design of a speech emotion recognition system, the choice of suitable features for speech representation, and the proper preparation of an emotional speech database for evaluating system performance are addressed. Expand
Real-Time Emotion Recognition from Natural Bodily Expressions in Child-Robot Interaction
TLDR
This paper investigates the spontaneous dimensional emotion prediction problem in a child-robot interaction scenario, and presents emotion elicitation, data acquisition, 3D skeletal representation, feature design and machine learning algorithms. Expand
Expressive speech synthesis using a concatenative synthesizer
TLDR
The highest recognition accuracies were achieved for sentences synthesized by using prosody and diphone inventory belonging to the same emotion, and anger was classified as inventory dominant and sadness and neutral as prosody dominant. Expand
Using Actor Portrayals to Systematically Study Multimodal Emotion Expression: The GEMEP Corpus
TLDR
It is argued in this paper that, in view of the needs of current research programs in this field, well-designed corpora of acted emotion portrayals can play a useful role. Expand
...
1
2
...