Audio-visual emotion recognition in video clips

F Noroozi, M Marjanovic, A Njegus… - IEEE Transactions …, 2017 - ieeexplore.ieee.org
This paper presents a multimodal emotion recognition system, which is based on the
analysis of audio and visual cues. From the audio channel, Mel-Frequency Cepstral …

Domain adversarial for acoustic emotion recognition

M Abdelwahab, C Busso - IEEE/ACM Transactions on Audio …, 2018 - ieeexplore.ieee.org
The performance of speech emotion recognition is affected by the differences in data
distributions between train (source domain) and test (target domain) sets used to build and …

Cross-corpus acoustic emotion recognition with multi-task learning: Seeking common ground while preserving differences

B Zhang, EM Provost, G Essl - IEEE Transactions on Affective …, 2017 - ieeexplore.ieee.org
There is growing interest in emotion recognition due to its potential in many applications.
However, a pervasive challenge is the presence of data variability caused by factors such as …

Recognizing emotion from singing and speaking using shared models

B Zhang, G Essl, EM Provost - 2015 international conference on …, 2015 - ieeexplore.ieee.org
Speech and song are two types of vocal communications that are closely related to each
other. While significant progress has been made in both speech and music emotion …

ISLA: Temporal segmentation and labeling for audio-visual emotion recognition

Y Kim, EM Provost - IEEE Transactions on affective computing, 2017 - ieeexplore.ieee.org
Emotion is an essential part of human interaction. Automatic emotion recognition can greatly
benefit human-centered interactive technology, since extracted emotion can be used to …

Improving the accuracy of automatic facial expression recognition in speaking subjects with deep learning

S Bursic, G Boccignone, A Ferrara, A D'Amelio… - Applied Sciences, 2020 - mdpi.com
When automatic facial expression recognition is applied to video sequences of speaking
subjects, the recognition accuracy has been noted to be lower than with video sequences of …

Report from the NSF future directions workshop, toward user-oriented agents: Research directions and challenges

M Eskenazi, T Zhao - arXiv preprint arXiv:2006.06026, 2020 - arxiv.org
This USER Workshop was convened with the goal of defining future research directions for
the burgeoning intelligent agent research community and to communicate them to the …

An investigation of partition-based and phonetically-aware acoustic features for continuous emotion prediction from speech

Z Huang, J Epps - IEEE Transactions on Affective Computing, 2018 - ieeexplore.ieee.org
Phonetic variability has long been considered a confounding factor for emotional speech
processing, so phonetic features have been rarely explored. However, surprisingly some …

Emotion spotting: Discovering regions of evidence in audio-visual emotion expressions

Y Kim, EM Provost - Proceedings of the 18th ACM international …, 2016 - dl.acm.org
Research has demonstrated that humans require different amounts of information, over time,
to accurately perceive emotion expressions. This varies as a function of emotion classes. For …

Facial expression recognition in the presence of speech using blind lexical compensation

S Mariooryad, C Busso - IEEE Transactions on Affective …, 2015 - ieeexplore.ieee.org
During spontaneous conversations the articulation process as well as the internal emotional
states influence the facial configurations. Inferring the conveyed emotions from the …