• Corpus ID: 160009690

Human Vocal Sentiment Analysis

  title={Human Vocal Sentiment Analysis},
  author={Andrew Huang and Puwei Bao},
In this paper, we use several techniques with conventional vocal feature extraction (MFCC, STFT), along with deep-learning approaches such as CNN, and also context-level analysis, by providing the textual data, and combining different approaches for improved emotion-level classification. [] Key Method We apply hyperparameter sweeps and data augmentation to improve performance. Finally, we see if a real-time approach is feasible, and can be readily integrated into existing systems.

Figures and Tables from this paper

Speech Emotion Based Sentiment Recognition using Deep Neural Networks
This research proposes a framework for acknowledging the passionate sections of conversation, independent of the semantic content, via the recognition of discourse feelings through deep learning techniques such as convolutional neural networks (CNNs) and long short-term memories (LSTMs).
Speech Emotion Recognition using MFCC and Hybrid Neural Networks
This paper proposes a hybrid neural network architecture based on Convolutional and Long Short-Term Memory (ConvLSTM) networks to capture spatial and sequential information of audio files.
MFCC and Machine Learning Based Speech Emotion Recognition Over TESS and IEMOCAP Datasets
Our proposed methodology involving MFCC computation along with support Vector machine is used to perform the task of Speech Emotion Recognition (SER) of collectively five emotions named Angry, Happy,
Sentiment Analysis of Meeting Room
  • Mert İleri, M. Turan
  • Computer Science
    2021 3rd International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA)
  • 2021
The artificial neural network was fed by training data and a classifier model was created, and it was seen that accuracy of model is about 85%.
Deteksi Emosi Wicara pada Media On-Demand menggunakan SVM dan LSTM
This paper describes the use of sound clips from HENLO, then conduct learning using Support Vector Machine (SVM) and Long Short-Term Memory (LSTM), finding the best strategy by training LSTMs first, then then feeding the model to SVM, with a data split strategy at 80:20 scale.
Human Voice Emotion Identification Using Prosodic and Spectral Feature Extraction Based on Deep Neural Networks
This paper proposes to identify the difference between the six essentials of human voice emotion, which will be generated using prosodic and spectral features extraction by utilizing Deep Neural Networks (DNNs).
An Enhanced Emotion Recognition Algorithm Using Pitch Correlogram, Deep Sparse Matrix Representation and Random Forest Classifier
This work presents an approach for text-independent and speaker-independent emotion recognition from speech in real application situations such as noisy and stressful talking conditions. We have
An Efficient Speech Emotion Recognition Using Ensemble Method of Supervised Classifiers
Based on a comparative analysis among six classifiers and other existing methods, it is revealed that Ensemble Method is one of the best alternative techniques for emotion recognition from speech.
Hierarchical classifier design for speech emotion recognition in the mixed-cultural environment
A two-level hierarchical engine has been designed to identify emotion from the speech of different cultural backgrounds, using a discriminative, multiclass SVM classifier trained with the emotional utterances of that particular corpus.
A Novel Framework Using Neutrosophy for Integrated Speech and Text Sentiment Analysis
A novel framework is proposed that performs sentiment analysis on audio files by calculating their Single-Valued Neutrosophic Sets (SVNS) and clustering them into positive-neutral-negative and combines these results with those obtained by performing sentimentAnalysis on the text files of those audio.


Utterance-Level Multimodal Sentiment Analysis
It is shown that multimodal sentiment analysis can be effectively performed, and that the joint use of visual, acoustic, and linguistic modalities can lead to error rate reductions of up to 10.5% as compared to the best performing individual modality.
Adieu features? End-to-end speech emotion recognition using a deep convolutional recurrent network
This paper proposes a solution to the problem of `context-aware' emotional relevant feature extraction, by combining Convolutional Neural Networks (CNNs) with LSTM networks, in order to automatically learn the best representation of the speech signal directly from the raw time representation.
Speech emotion recognition using deep neural network and extreme learning machine
The experimental results demonstrate that the proposed approach effectively learns emotional information from low-level features and leads to 20% relative accuracy improvement compared to the state of the art approaches.
Emotion Recognition in Speech using Cross-Modal Transfer in the Wild
A strong teacher network for facial emotion recognition that achieves the state of the art on a standard benchmark is developed and it is shown that the speech emotion embedding can be used for speech emotion recognition on external benchmark datasets.
High-level feature representation using recurrent neural network for speech emotion recognition
This paper presents a speech emotion recognition system using a recurrent neural network (RNN) model trained by an efficient learning algorithm. The proposed system takes into account the long-range
Speech Emotion Recognition Using Deep Convolutional Neural Network and Discriminant Temporal Pyramid Matching
This paper explores how to utilize a DCNN to bridge the affective gap in speech signals, and finds that the DCNN model pretrained for image applications performs reasonably good in affective speech feature extraction.
Speech emotion recognition combining acoustic features and linguistic information in a hybrid support vector machine-belief network architecture
A novel approach to the combination of acoustic features and language information for a most robust automatic recognition of a speaker's emotion by applying belief network based spotting for emotional key-phrases is introduced.
Emotion recognition by speech signals
Experimental results showed that pitch and energy were the most important factors in emotion recognition, and compared emotion recognition performance of various classifiers: SVM, linear discriminant analysis (LDA), QDA and hidden Markov model (HMM).
Detection and Analysis of Human Emotions through Voice and Speech Pattern Processing
This paper presents an algorithmic approach for detection and analysis of human emotions with the help of voice and speech processing with the objective of incorporation with futuristic artificial intelligence systems for improving human-computer interactions.
The History Began from AlexNet: A Comprehensive Survey on Deep Learning Approaches
This report presents a brief survey on development of DL approaches, including Deep Neural Network (DNN), Convolutional neural network (CNN), Recurrent Neural network (RNN) including Long Short Term Memory (LSTM) and Gated Recurrent Units (GRU), Auto-Encoder (AE), Deep Belief Network (DBN), Generative Adversarial Network (GAN), and Deep Reinforcement Learning (DRL).