Corpus ID: 235458066

Automatic Analysis of the Emotional Content of Speech in Daylong Child-Centered Recordings from a Neonatal Intensive Care Unit

  title={Automatic Analysis of the Emotional Content of Speech in Daylong Child-Centered Recordings from a Neonatal Intensive Care Unit},
  author={Einari Vaaras and S. Ahlqvist-Bj{\"o}rkroth and Konstantinos Drossos and O. R{\"a}s{\"a}nen},
Researchers have recently started to study how the emotional speech heard by young infants can affect their developmental outcomes. As a part of this research, hundreds of hours of daylong recordings from preterm infants’ audio environments were collected from two hospitals in Finland and Estonia in the context of so-called APPLE study. In order to analyze the emotional content of speech in such a massive dataset, an automatic speech emotion recognition (SER) system is required. However, there… Expand

Figures and Tables from this paper


LSSED: A Large-Scale Dataset and Benchmark for Speech Emotion Recognition
A challenging large-scale english speech emotion dataset, which has data collected from 820 subjects to simulate real- world distribution, and some pre-trained models based on LSSED, which can not only promote the development of speech emotion recognition, but can also be transferred to related downstream tasks such as mental health analysis where data is extremely difficult to collect. Expand
Active Learning for Speech Emotion Recognition Using Deep Neural Network
  • M. Abdelwahab, C. Busso
  • Computer Science
  • 2019 8th International Conference on Affective Computing and Intelligent Interaction (ACII)
  • 2019
This study explores practical solutions to train DNNs for speech emotion recognition with limited resources by using active learning (AL), assuming that data without emotional labels from a new domain are available and one has resources to select a limited number of recordings to be annotated with emotional labels. Expand
A database of German emotional speech
A database of emotional speech that was evaluated in a perception test regarding the recognisability of emotions and their naturalness and can be accessed by the public via the internet. Expand
A thorough evaluation of the Language Environment Analysis (LENA) system.
Whether LENAⓇ results are accurate enough for a given research, educational, or clinical application depends largely on the specifics at hand, and a set of recommendations is concluded to help researchers make this determination for their goals. Expand
Universum Autoencoder-Based Domain Adaptation for Speech Emotion Recognition
This letter proposes a novel unsupervised domain adaptation model, called Universum autoencoders, to improve the performance of the systems evaluated in mismatched training and test conditions and demonstrates the effectiveness of the proposed method when compared to other domain adaptation methods. Expand
The Automatic Recognition of Emotions in Speech
The subject area of this chapter is not emotions in some narrow sense but in a wider sense encompassing emotion-related states such as moods, attitudes, or interpersonal stances as well. Expand
Unsupervised learning in cross-corpus acoustic emotion recognition
It is shown that adding unlabeled emotional speech to agglomerated multi-corpus training sets can enhance recognition performance even in a challenging cross- Corpus setting, and that the expected gain by adding unl Isabeled data on average is approximately half the one achieved by additional manually labeled data in leave-one-corpsus-out validation. Expand
The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing
A basic standard acoustic parameter set for various areas of automatic voice analysis, such as paralinguistic or clinical speech analysis, is proposed and intended to provide a common baseline for evaluation of future research and eliminate differences caused by varying parameter sets or even different implementations of the same parameters. Expand
Active Learning for Speech Emotion Recognition Using Conditional Random Fields
  • Ziping Zhao, Xirong Ma
  • Computer Science
  • 2013 14th ACIS International Conference on Software Engineering, Artificial Intelligence, Networking and Parallel/Distributed Computing
  • 2013
Experiments show that for most of the cases considered, active selection strategies when recognizing speech emotion are as good as or exceed the performance of random data selection. Expand
Unsupervised Adversarial Domain Adaptation for Cross-Lingual Speech Emotion Recognition
The proposed GAN-based model for multilingual SER is designed in such a way that the language invariant representations can be learned without requiring target-language data labels and can significantly improve the baseline cross-lingual SER performance. Expand