A systematic review on affective computing: Emotion models, databases, and recent advances

Y Wang, W Song, W Tao, A Liotta, D Yang, X Li, S Gao… - Information …, 2022 - Elsevier
Affective computing conjoins the research topics of emotion recognition and sentiment
analysis, and can be realized with unimodal or multimodal data, consisting primarily of …

Deep learning-based multimodal emotion recognition from audio, visual, and text modalities: A systematic review of recent advancements and future prospects

S Zhang, Y Yang, C Chen, X Zhang, Q Leng… - Expert Systems with …, 2024 - Elsevier
Emotion recognition has recently attracted extensive interest due to its significant
applications to human–computer interaction. The expression of human emotion depends on …

Multimodal emotion recognition using deep learning

SMSA Abdullah, SYA Ameen, MAM Sadeeq… - Journal of Applied …, 2021 - jastt.org
New research into human-computer interaction seeks to consider the consumer's emotional
status to provide a seamless human-computer interface. This would make it possible for …

[HTML][HTML] Survey on bimodal speech emotion recognition from acoustic and linguistic information fusion

BT Atmaja, A Sasou, M Akagi - Speech Communication, 2022 - Elsevier
Speech emotion recognition (SER) is traditionally performed using merely acoustic
information. Acoustic features, commonly are extracted per frame, are mapped into emotion …

Head fusion: Improving the accuracy and robustness of speech emotion recognition on the IEMOCAP and RAVDESS dataset

M Xu, F Zhang, W Zhang - IEEE Access, 2021 - ieeexplore.ieee.org
Speech Emotion Recognition (SER) refers to the use of machines to recognize the emotions
of a speaker from his (or her) speech. SER benefits Human-Computer Interaction (HCI). But …

Emotional speech recognition using deep neural networks

L Trinh Van, T Dao Thi Le, T Le Xuan, E Castelli - Sensors, 2022 - mdpi.com
The expression of emotions in human communication plays a very important role in the
information that needs to be conveyed to the partner. The forms of expression of human …

Speech emotion recognition with multiscale area attention and data augmentation

M Xu, F Zhang, X Cui, W Zhang - ICASSP 2021-2021 IEEE …, 2021 - ieeexplore.ieee.org
In Speech Emotion Recognition (SER), emotional characteristics often appear in diverse
forms of energy patterns in spectrograms. Typical attention neural network classifiers of SER …

M3GAT: A multi-modal, multi-task interactive graph attention network for conversational sentiment analysis and emotion recognition

Y Zhang, A Jia, B Wang, P Zhang, D Zhao, P Li… - ACM Transactions on …, 2023 - dl.acm.org
Sentiment and emotion, which correspond to long-term and short-lived human feelings, are
closely linked to each other, leading to the fact that sentiment analysis and emotion …

Self-supervised contrastive cross-modality representation learning for spoken question answering

C You, N Chen, Y Zou - arXiv preprint arXiv:2109.03381, 2021 - arxiv.org
Spoken question answering (SQA) requires fine-grained understanding of both spoken
documents and questions for the optimal answer prediction. In this paper, we propose novel …

Efficient long-distance latent relation-aware graph neural network for multi-modal emotion recognition in conversations

Y Shou, W Ai, J Du, T Meng, H Liu, N Yin - arXiv preprint arXiv:2407.00119, 2024 - arxiv.org
The task of multi-modal emotion recognition in conversation (MERC) aims to analyze the
genuine emotional state of each utterance based on the multi-modal information in the …