Summary of MuSe 2020: Multimodal Sentiment Analysis, Emotion-target Engagement and Trustworthiness Detection in Real-life Media
@article{Stappen2020SummaryOM, title={Summary of MuSe 2020: Multimodal Sentiment Analysis, Emotion-target Engagement and Trustworthiness Detection in Real-life Media}, author={Lukas Stappen and Bj{\"o}rn W. Schuller and Iulia Lefter and E. Cambria and Yiannis Kompatsiaris}, journal={Proceedings of the 28th ACM International Conference on Multimedia}, year={2020} }
The first Multimodal Sentiment Analysis in Real-life Media (MuSe) 2020 was a Challenge-based Workshop held in conjunction with ACM Multimedia'20. It addresses three distinct 'in-the-wild` Sub-challenges: sentiment/ emotion recognition (MuSe-Wild), emotion-target engagement (MuSe-Target) and trustworthiness detection (MuSe-Trust). A large multimedia dataset MuSe-CaR was used, which was specifically designed with the intention of improving machine understanding approaches of how sentiment (e.g…
10 Citations
MuSe 2021 Challenge: Multimodal Emotion, Sentiment, Physiological-Emotion, and Stress Detection
- Computer ScienceACM Multimedia
- 2021
The motivation, the sub-challenges, the challenge conditions, the participation, the Participation, and the most successful approaches are described.
The MuSe 2021 Multimodal Sentiment Analysis Challenge: Sentiment, Emotion, Physiological-Emotion, and Stress
- Computer ScienceMuSe @ ACM Multimedia
- 2021
This paper utilises the MuSe-CaR dataset focusing on user-generated reviews and introduces the Ulm-TSST dataset, which displays people in stressful depositions, and provides detail on the state-of-the-art feature sets extracted from these datasets for utilisation by the baseline model, a Long Short-Term Memory-Recurrent Neural Network.
The Multimodal Sentiment Analysis in Car Reviews (MuSe-CaR) Dataset: Collection, Insights and Improvements
- Computer ScienceIEEE Transactions on Affective Computing
- 2021
This contribution presents MuSe-CaR, a first of its kind multimodal dataset that focused on the tasks of emotion, emotion-target engagement, and trustworthiness recognition by means of comprehensively integrating the audio-visual and language modalities and gives a thorough overview of the dataset in terms of collection and annotation.
Sentiment Analysis and Topic Recognition in Video Transcriptions
- Computer ScienceIEEE Intelligent Systems
- 2021
This article uses SenticNet to extract natural language concepts and fine-tune several feature types on a subset of MuSe-CAR to explore the content of a video as well as learning to predict emotional valence, arousal, and speaker topic classes.
An Estimation of Online Video User Engagement From Features of Time- and Value-Continuous, Dimensional Emotions
- Computer ScienceFrontiers in Computer Science
- 2022
It is demonstrated that smaller boundary ranges and fluctuations for arousal lead to an increase in user engagement, and an effective combination of features is outlined for approaches aiming to automatically predict several user engagement indicators.
MuSe-Toolbox: The Multimodal Sentiment Analysis Continuous Annotation Fusion and Discrete Class Transformation Toolbox
- Computer ScienceMuSe @ ACM Multimedia
- 2021
This work introduces the MuSe-Toolbox - a Python-based open-source toolkit for creating a variety of continuous and discrete emotion gold standards and proposes the novel Rater Aligned Annotation Weighting (RAAW), which aligns the annotations in a translation-invariant way before weighting and fusing them based on the inter-rater agreements between the annotations.
A survey on sentiment analysis methods, applications, and challenges
- Computer ScienceArtificial Intelligence Review
- 2022
An overview of the method for completing this task as well as the applications of sentiment analysis is discussed, which evaluates, compares, and investigates the approaches used to gain a comprehensive understanding of their advantages and disadvantages.
Temporal Graph Convolutional Network for Multimodal Sentiment Analysis
- Computer ScienceICMI
- 2021
This paper uses positional encoding by interleaving sine and cosine embedding to encode the positions of the segments in the utterances into their features and creates an attention mechanism corresponding to the segments to capture the sentiment-related ones and obtain the unified embeddings of utterances.
The INTERSPEECH 2021 Computational Paralinguistics Challenge: COVID-19 Cough, COVID-19 Speech, Escalation & Primates
- Computer ScienceInterspeech
- 2021
The Sub-Challenges, baseline feature extraction, and classifiers based on the 'usual' COMPARE and BoAW features as well as deep unsupervised representation learning using the AuDeep toolkit, and deep feature extraction from pre-trained CNNs using the Deep Spectrum toolkit are described.
An Estimation of Online Video User Engagement from Features of Continuous Emotions
- Computer ScienceArXiv
- 2021
It is demonstrated that smaller boundary ranges and fluctuations for arousal lead to an increase in user engagement, and an effective combination of features is outlined for approaches aiming to automatically predict several user engagement indicators.
References
SHOWING 1-8 OF 8 REFERENCES
AVEC 2013: the continuous audio/visual emotion and depression recognition challenge
- PsychologyAVEC@ACM Multimedia
- 2013
The third Audio-Visual Emotion recognition Challenge (AVEC 2013) has two goals logically organised as sub-challenges: the first is to predict the continuous values of the affective dimensions valence and arousal at each moment in time, and the second is to Predict the value of a single depression indicator for each recording in the dataset.
AVEC 2017: Real-life Depression, and Affect Recognition Workshop and Challenge
- Computer ScienceAVEC@ACM Multimedia
- 2017
This paper presents the novelties introduced this year, the challenge guidelines, the data used, and the performance of the baseline system on the two proposed tasks: dimensional emotion recognition (time and value-continuous), and dimensional depression estimation (value-Continuous).
Integrating Multimodal Information in Large Pretrained Transformers
- Computer ScienceACL
- 2020
Fine-tuning MAG-BERT and MAG-XLNet significantly boosts the sentiment analysis performance over previous baselines as well as language-only fine- Tuning of BERT and XLNet.
The INTERSPEECH 2020 Computational Paralinguistics Challenge: Elderly Emotion, Breathing & Masks
- Computer ScienceINTERSPEECH
- 2020
The Sub-Challenges, baseline feature extraction, and classifiers based on the ‘usual’ COMPARE and BoAW features as well as deep unsupervised representation learning using the AUDEEP toolkit, and deep feature extraction from pre-trained CNNs using the DEEP SPECTRUM toolkit are described.
Erik Cambria, and Ioannis Kompatsiaris. 2020. MuSe 2020 Challenge and Workshop: Multimodal Sentiment Analysis, Emotion-target Engagement and Trustworthiness Detection in Real-life Media
- 1st International Multimodal Sentiment Analysis in Real-life Media Challenge and Workshop, co-located with the 28th ACM International Conference on Multimedia
Challenge and Workshop: Multimodal Sentiment Analysis, Emotion-target Engagement and Trustworthiness Detection in Real-life Media
- MuSe
- 2020