IEMOCAP: interactive emotional dyadic motion capture database

@article{Busso2008IEMOCAPIE,
  title={IEMOCAP: interactive emotional dyadic motion capture database},
  author={C. Busso and M. Bulut and Chi-Chun Lee and Ebrahim Kazemzadeh and E. Provost and Samuel Kim and J. N. Chang and Sungbok Lee and Shrikanth S. Narayanan},
  journal={Language Resources and Evaluation},
  year={2008},
  volume={42},
  pages={335-359}
}
Since emotions are expressed through a combination of verbal and non-verbal channels, a joint analysis of speech and gestures is required to understand expressive human communication. [...] Key Result The detailed motion capture information, the interactive setting to elicit authentic emotions, and the size of the database make this corpus a valuable addition to the existing databases in the community for the study and modeling of multimodal and expressive human communication.Expand
RAMAS: Russian Multimodal Corpus of Dyadic Interaction for Affective Computing
TLDR
The Russian Acted Multimodal Affective Set (RAMAS) − the first multimodal corpus in Russian language is collected and the annotation process, inter-rater agreement analysis and the comparison between self-reports and received annotations are presented. Expand
RAMAS: Russian Multimodal Corpus of Dyadic Interaction for studying emotion recognition
Emotion expression encompasses various types of information, including face and eye movement, voice and body motion. Most of the studies in automated affective recognition use faces as stimuli, lessExpand
Acquisition of a 3D audio-visual corpus of affective speech
TLDR
This work presents a new audio-visual corpus for possibly the two most important modalities used by humans to communicate their emotional states, namely speech and facial expression in the form of dense dynamic 3D face geometries and introduces an acquisition setup for labeling the data with very little manual effort. Expand
The JESTKOD database: an affective multimodal database of dyadic interactions
TLDR
The JESTKOD database is introduced, which consists of speech and full-body motion capture data recordings in dyadic interaction setting under agreement and disagreement scenarios and provides a valuable asset to investigate gesture and speech towards designing more natural and affective human–computer interaction systems. Expand
Recording audio-visual emotional databases from actors : a closer look
Research on human emotional behavior, and the development of automatic emotion recognition and animation systems, rely heavily on appropriate audio-visual databases of expressive human speech,Expand
A 3-D Audio-Visual Corpus of Affective Communication
TLDR
This work presents a new audio-visual corpus for possibly the two most important modalities used by humans to communicate their emotional states, namely speech and facial expression in the form of dense dynamic 3-D face geometries. Expand
The USC CreativeIT database of multimodal dyadic interactions: from speech and full body motion capture to continuous emotional annotations
TLDR
The carefully engineered data collection, the improvisation design to elicit natural emotions and expressive speech and body language, as well as the well-developed annotation processes provide a gateway to study and model various aspects of theatrical performance, expressive behaviors and human communication and interaction. Expand
Emotion recognition from embedded bodily expressions and speech during dyadic interactions
TLDR
This work studies the problem of emotion recognition from bodily expressions and speech during dyadic ( person-person) interactions in a real kitchen instrumented with ambient cameras and microphones and presents a human-validated dataset that contains 224 high-resolution, multi-view video clips and audio recordings of emotionally charged interactions between eight couples of actors. Expand
MSP-Face Corpus: A Natural Audiovisual Emotional Database
TLDR
This study presents the MSP-Face database, a natural audiovisual database obtained from video-sharing websites, where multiple individuals discuss various topics expressing their opinions and experiences, offering the perfect infrastructure to explore semi-supervised and unsupervised machine-learning algorithms on natural emotional videos. Expand
MSP-AVATAR corpus: Motion capture recordings to study the role of discourse functions in the design of intelligent virtual agents
  • Najmeh Sadoughi, Y. Liu, C. Busso
  • Computer Science
  • 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG)
  • 2015
TLDR
The MSP-AVATAR data is introduced, a new multimedia corpus designed to explore the relationship between discourse functions, speech and nonverbal behaviors, and is suitable for rule-based and speech-based generation of body, hand and facial behaviors for IVAs. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 75 REFERENCES
Recording audio-visual emotional databases from actors : a closer look
Research on human emotional behavior, and the development of automatic emotion recognition and animation systems, rely heavily on appropriate audio-visual databases of expressive human speech,Expand
Modeling naturalistic affective states via facial and vocal expressions recognition
TLDR
A multi-cue, dynamic approach in naturalistic video sequences, deployed on an audiovisual database which was recorded simulating human-human discourse and, therefore, contains less extreme expressivity and subtle variations of a number of emotion labels. Expand
The Vera am Mittag German audio-visual emotional speech database
TLDR
This contribution presents a recently collected database of spontaneous emotional speech in German which is being made available to the research community and provides emotion labels for a great part of the data. Expand
EmoTV 1 : Annotation of Real-life Emotions for the Specification of Multimodal Affective Interfaces
  • 2005
The development of future multimodal affective interfaces such as believable Embodied Conversational Agents requires to model relations between natural emotions and multimodal behaviors in variousExpand
Interrelation Between Speech and Facial Gestures in Emotional Utterances: A Single Subject Study
TLDR
The results suggest that emotional content affect the relationship between facial gestures and speech, and principal component analysis (PCA) shows that the audiovisual mapping parameters are grouped in a smaller subspace, which suggests that there is an emotion-dependent structure that is preserved from across sentences. Expand
Analysis of emotion recognition using facial expressions, speech and multimodal information
TLDR
Results reveal that the system based on facial expression gave better performance than the systembased on just acoustic information for the emotions considered, and that when these two modalities are fused, the performance and the robustness of the emotion recognition system improve measurably. Expand
Joint Analysis of the Emotional Fingerprint in the Face and Speech: A single subject study
TLDR
The results presented here support the hypothesis that facial expression and prosodic speech tend to have a stronger emotional modulation when the vocal tract is physically constrained by the articulation to convey other linguistic communicative goals. Expand
Rigid Head Motion in Expressive Speech Animation: Analysis and Synthesis
TLDR
The results suggest that appropriate head motion not only significantly improves the naturalness of the animation but can also be used to enhance the emotional content of theAnimation to effectively engage the users. Expand
Scripted dialogs versus improvisation: lessons learned about emotional elicitation techniques from the IEMOCAP database
TLDR
Analysis of two of the most appealing elicitation approaches: improvisation, and scripted dialogs indicates that spontaneous sessions have higher levels of disfluencies and overlapped speech, and the emotional content seems to be more intense than in scripted sessions. Expand
Interplay between linguistic and affective goals in facial expression during emotional utterances
Communicative goals are simultaneously expressed through gestures and speech to convey messages enriched with valuable verbal and non-verbal clues. This paper analyzes and quantifies how linguisticExpand
...
1
2
3
4
5
...