Cross-corpus acoustic emotion recognition: Variances and strategies

B Schuller, B Vlasenko, F Eyben… - IEEE Transactions …, 2010 - ieeexplore.ieee.org
As the recognition of emotion from speech has matured to a degree where it becomes
applicable in real-life settings, it is time for a realistic view on obtainable performances. Most …

Combining long short-term memory and dynamic bayesian networks for incremental emotion-sensitive artificial listening

M Wöllmer, B Schuller, F Eyben… - IEEE Journal of selected …, 2010 - ieeexplore.ieee.org
The automatic estimation of human affect from the speech signal is an important step
towards making virtual agents more natural and human-like. In this paper, we present a …

[PDF][PDF] Audio-visual feature selection and reduction for emotion classification.

S Haq, PJB Jackson, JD Edge - AVSP, 2008 - isca-archive.org
Recognition of expressed emotion from speech and facial gestures was investigated in
experiments on an audio-visual emotional database. A total of 106 audio and 240 visual …

On-line emotion recognition in a 3-D activation-valence-time continuum using acoustic and linguistic cues

F Eyben, M Wöllmer, A Graves, B Schuller… - Journal on Multimodal …, 2010 - Springer
For many applications of emotion recognition, such as virtual agents, the system must select
responses while the user is speaking. This requires reliable on-line recognition of the user's …

Speech emotion recognition research: an analysis of research focus

MB Mustafa, MAM Yusoof, ZM Don… - International Journal of …, 2018 - Springer
This article analyses research in speech emotion recognition (“SER”) from 2006 to 2017 in
order to identify the current focus of research, and areas in which research is lacking. The …

[PDF][PDF] Towards temporal modelling of categorical speech emotion recognition

W Han, H Ruan, X Chen, Z Wang, H Li, B Schuller - 2018 - opus.bibliothek.uni-augsburg.de
To model the categorical speech emotion recognition task in a temporal manner, the first
challenge arising is how to transfer the categorical label for each utterance into a label …

Pretext tasks selection for multitask self-supervised audio representation learning

S Zaiem, T Parcollet, S Essid… - IEEE Journal of Selected …, 2022 - ieeexplore.ieee.org
Through solving pretext tasks, self-supervised learning leverages unlabeled data to extract
useful latent representations replacing traditional input features in the downstream task. In …

Iterative feature normalization scheme for automatic emotion detection from speech

C Busso, S Mariooryad, A Metallinou… - IEEE transactions on …, 2013 - ieeexplore.ieee.org
The externalization of emotion is intrinsically speaker-dependent. A robust emotion
recognition system should be able to compensate for these differences across speakers. A …

Recognizing affect from linguistic information in 3D continuous space

B Schuller - IEEE Transactions on Affective computing, 2011 - ieeexplore.ieee.org
Most research efforts dealing with recognition of emotion-related states from the human
speech signal concentrate on acoustic analysis. However, the last decade's research results …

A multitask approach to continuous five-dimensional affect sensing in natural speech

F Eyben, M Wöllmer, B Schuller - ACM Transactions on Interactive …, 2012 - dl.acm.org
Automatic affect recognition is important for the ability of future technical systems to interact
with us socially in an intelligent way by understanding our current affective state. In recent …