Corpus ID: 20160536

Multimodal Emotion Recognition for One-Minute-Gradual Emotion Challenge

@article{Zheng2018MultimodalER,
  title={Multimodal Emotion Recognition for One-Minute-Gradual Emotion Challenge},
  author={Ziqi Zheng and Chenjie Cao and Xingwei Chen and Guoqiang Xu},
  journal={ArXiv},
  year={2018},
  volume={abs/1805.01060}
}
The continuous dimensional emotion modelled by arousal and valence can depict complex changes of emotions. In this paper, we present our works on arousal and valence predictions for One-Minute-Gradual (OMG) Emotion Challenge. Multimodal representations are first extracted from videos using a variety of acoustic, video and textual models and support vector machine (SVM) is then used for fusion of multimodal signals to make final predictions. Our solution achieves Concordant Correlation… Expand
A Personalized Affective Memory Neural Model for Improving Emotion Recognition
TLDR
A neural model based on a conditional adversarial autoencoder to learn how to represent and edit general emotion expressions is presented and Grow-When-Required networks are proposed as personalized affective memories to learn individualized aspects of emotion expressions. Expand
Multimodal Deep Models for Predicting Affective Responses Evoked by Movies
The goal of this study is to develop and analyze multimodal models for predicting experienced affective responses of viewers watching movie clips. We develop hybrid multimodal prediction models basedExpand
AttendAffectNet: Self-Attention based Networks for Predicting Affective Responses from Movies
TLDR
The results show that applying the self-attention mechanism on the different audio-visual features, rather than in the time domain, is more effective for emotion prediction, and is also proven to outperform many state-of-the-art models for emotion Prediction. Expand
Arousal and Valence Estimation for Visual Non-Intrusive Stress Monitoring
TLDR
A deep learning-based psychological stress level estimation approach to identify the region where the emotional state of the operator projects in the space defined by the latent dimensional emotions of arousal and valence. Expand
Exploiting Multi-CNN Features in CNN-RNN Based Dimensional Emotion Recognition on the OMG in-the-Wild Dataset
TLDR
This paper presents a novel CNN-RNN based approach, which exploits multiple CNN features for dimensional emotion recognition in-the-wild, utilizing the One-Minute Gradual-Emotion (OMG) dataset, and shows that arousal estimation is greatly improved when low-level features are combined with high-level ones. Expand
Affect Analysis in-the-wild: Valence-Arousal, Expressions, Action Units and a Unified Framework
TLDR
A novel multi-task and holistic framework is presented which is able to jointly learn and effectively generalize and perform affect recognition over all existing in-the-wild databases. Expand
A Deep Neural Model Of Emotion Appraisal
TLDR
A deep neural model is proposed which is designed in the light of different aspects of developmental learning of emotional concepts to provide an integrated solution for internal and external emotion appraisal and is able to generate internal emotional concepts that evolve through time. Expand
The OMG-Empathy Dataset: Evaluating the Impact of Affective Behavior in Storytelling
TLDR
This paper proposes a novel dataset composed of dyadic interactions designed, collected and annotated with a focus on measuring the affective impact that eight different stories have on the listener, and proposes different evaluation protocols and a baseline that encourages participation in the advancement of the field of artificial empathy and emotion contagion. Expand
MIMAMO Net: Integrating Micro- and Macro-motion for Video Emotion Recognition
TLDR
This paper proposes to combine micro- and macro-motion features to improve video emotion recognition with a two-stream recurrent network, named MIMAMO (Micro-Macro-Motion) Net, and achieves state of the art performance on two video emotion datasets, the OMG emotion dataset and the Aff-Wild dataset. Expand
Privileged Information for Modeling Affect In The Wild
TLDR
The proposed privileged information framework enables affect models to be trained across multiple modalities available in a lab setting and ignore modalities that are not available in the wild with no significant drop in their modeling performance. Expand
...
1
2
...

References

SHOWING 1-10 OF 32 REFERENCES
Continuous Multimodal Emotion Prediction Based on Long Short Term Memory Recurrent Neural Network
TLDR
This paper presents the automatic prediction of dimensional emotional state for Audio-Visual Emotion Challenge (AVEC 2017), which uses multi- features and fusion across all available modalities and achieves very promising results with Concordant Correlation Coefficient. Expand
Multimodal Multi-task Learning for Dimensional and Continuous Emotion Recognition
TLDR
This paper presents the effort for the Affect Subtask in the Audio/Visual Emotion Challenge (AVEC) 2017, which requires participants to perform continuous emotion prediction on three affective dimensions: Arousal, Valence and Likability based on the audiovisual signals and highlights three aspects of the solutions. Expand
The OMG-Emotion Behavior Dataset
TLDR
This paper proposes a novel multimodal corpus for emotion expression recognition, which uses gradual annotations with a focus on contextual emotion expressions and provides an experimental protocol and a series of unimodal baseline experiments which can be used to evaluate deep and recurrent neural models in a fair and standard manner. Expand
Modeling multimodal cues in a deep learning-based framework for emotion recognition in the wild
TLDR
A fusion network that merges cues from the different modalities in one representation is proposed that outperforms the challenge baselines and achieves an accuracy of 50.39 % and 49.92 % respectively on the validation and the testing data. Expand
Video-based emotion recognition using CNN-RNN and C3D hybrid networks
TLDR
Extensive experiments show that combining RNN and C3D together can improve video-based emotion recognition noticeably, and are presented to the EmotiW 2016 Challenge. Expand
HoloNet: towards robust emotion recognition in the wild
TLDR
HoloNet, a well-designed Convolutional Neural Network (CNN) architecture regarding the authors' submissions to the video based sub-challenge of the Emotion Recognition in the Wild (EmotiW) 2016 challenge, achieves competitive results in this year’s video based emotion recognition sub-Challenge. Expand
Hierarchical committee of deep convolutional neural networks for robust facial expression recognition
TLDR
This paper builds a hierarchical architecture of the committee with exponentially-weighted decision fusion of deep CNNs for robust facial expression recognition for the third Emotion Recognition in the Wild (EmotiW2015) challenge. Expand
Specificity of facial emotion recognition impairments in patients with multi-episode schizophrenia
TLDR
The hypothesis that facial emotion recognition impairment is a relatively distinct entity within the domain of cognitive dysfunction in schizophrenia is supported. Expand
Convolutional Neural Networks for Sentence Classification
TLDR
The CNN models discussed herein improve upon the state of the art on 4 out of 7 tasks, which include sentiment analysis and question classification, and are proposed to allow for the use of both task-specific and static vectors. Expand
FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition
  • Hui Ding, S. Zhou, R. Chellappa
  • Computer Science
  • 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017)
  • 2017
TLDR
A novel idea to train an expression recognition network based on static images, using a new distribution function to model the high-level neurons of the expression network and achieves better results than state-of-the-art. Expand
...
1
2
3
4
...