Multimodal speech emotion recognition using audio and text

S Yoon, S Byun, K Jung - 2018 IEEE spoken language …, 2018 - ieeexplore.ieee.org
Speech emotion recognition is a challenging task, and extensive reliance has been placed
on models that use audio features in building well-performing classifiers. In this paper, we …

Automatic speech emotion recognition using recurrent neural networks with local attention

S Mirsamadi, E Barsoum… - 2017 IEEE International …, 2017 - ieeexplore.ieee.org
Automatic emotion recognition from speech is a challenging task which relies heavily on the
effectiveness of the speech features used for classification. In this work, we study the use of …

Building naturalistic emotionally balanced speech corpus by retrieving emotional speech from existing podcast recordings

R Lotfian, C Busso - IEEE Transactions on Affective Computing, 2017 - ieeexplore.ieee.org
The lack of a large, natural emotional database is one of the key barriers to translate results
on speech emotion recognition in controlled conditions into real-life applications. Collecting …

MSP-IMPROV: An acted corpus of dyadic interactions to study emotion perception

C Busso, S Parthasarathy, A Burmania… - IEEE Transactions …, 2016 - ieeexplore.ieee.org
We present the MSP-IMPROV corpus, a multimodal emotional database, where the goal is
to have control over lexical content and emotion while also promoting naturalness in the …

Domain invariant feature learning for speaker-independent speech emotion recognition

C Lu, Y Zong, W Zheng, Y Li, C Tang… - … /ACM Transactions on …, 2022 - ieeexplore.ieee.org
In this paper, we propose a novel domain invariant feature learning (DIFL) method to deal
with speaker-independent speech emotion recognition (SER). The basic idea of DIFL is to …

Domain adversarial for acoustic emotion recognition

M Abdelwahab, C Busso - IEEE/ACM Transactions on Audio …, 2018 - ieeexplore.ieee.org
The performance of speech emotion recognition is affected by the differences in data
distributions between train (source domain) and test (target domain) sets used to build and …

Speech emotion recognition using multi-hop attention mechanism

S Yoon, S Byun, S Dey, K Jung - ICASSP 2019-2019 IEEE …, 2019 - ieeexplore.ieee.org
In this paper, we are interested in exploiting textual and acoustic data of an utterance for the
speech emotion classification task. The baseline approach models the information from …

Human–computer interaction with a real-time speech emotion recognition with ensembling techniques 1D convolution neural network and attention

W Alsabhan - Sensors, 2023 - mdpi.com
Emotions have a crucial function in the mental existence of humans. They are vital for
identifying a person's behaviour and mental condition. Speech Emotion Recognition (SER) …

The ordinal nature of emotions: An emerging approach

GN Yannakakis, R Cowie… - IEEE Transactions on …, 2018 - ieeexplore.ieee.org
Computational representation of everyday emotional states is a challenging task and,
arguably, one of the most fundamental for affective computing. Standard practice in emotion …

Multi-head attention fusion networks for multi-modal speech emotion recognition

J Zhang, L Xing, Z Tan, H Wang, K Wang - Computers & Industrial …, 2022 - Elsevier
Multi-modal speech emotion recognition is a study to predict emotion categories by
combining speech data with other types of data, such as video, speech text transcription …