Learning to listen: Modeling non-deterministic dyadic facial motion

E Ng, H Joo, L Hu, H Li, T Darrell… - Proceedings of the …, 2022 - openaccess.thecvf.com
We present a framework for modeling interactional communication in dyadic conversations:
given multimodal inputs of a speaker, we autoregressively output multiple possibilities of …

Can language models learn to listen?

E Ng, S Subramanian, D Klein… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present a framework for generating appropriate facial responses from a listener in
dyadic social interactions based on the speaker's words. Given an input transcription of the …

ZeroEGGS: Zero‐shot Example‐based Gesture Generation from Speech

S Ghorbani, Y Ferstl, D Holden, NF Troje… - Computer Graphics …, 2023 - Wiley Online Library
We present ZeroEGGS, a neural network framework for speech‐driven gesture generation
with zero‐shot style control by example. This means style can be controlled via only a short …

What affects the usage of artificial conversational agents? An agent personality and love theory perspective

D Pal, V Vanijja, H Thapliyal, X Zhang - Computers in Human Behavior, 2023 - Elsevier
Artificial-intelligence (AI) powered conversational (CAI) agents have been growing in
popularity. Like human personality, CAI agent personality can also impact the relationship …

Robotic mental well-being coaches for the workplace: An in-the-wild study on form

M Spitale, M Axelsson, H Gunes - Proceedings of the 2023 ACM/IEEE …, 2023 - dl.acm.org
The World Health Organization recommends that employers take action to protect and
promote mental well-being at work. However, the extent to which these recommended …

ExpressGesture: Expressive gesture generation from speech through database matching

Y Ferstl, M Neff, R McDonnell - Computer Animation and Virtual …, 2021 - Wiley Online Library
Co‐speech gestures are a vital ingredient in making virtual agents more human‐like and
engaging. Automatically generated gestures based on speech‐input often lack realistic and …

The MuSe 2024 multimodal sentiment analysis challenge: Social perception and humor recognition

S Amiriparian, L Christ, A Kathan, M Gerczuk… - Proceedings of the 5th …, 2024 - dl.acm.org
The Multimodal Sentiment Analysis Challenge (MuSe) 2024 addresses two contemporary
multimodal affect and sentiment analysis problems: In the Social Perception Sub-Challenge …

Affective faces for goal-driven dyadic communication

S Geng, R Teotia, P Tendulkar, S Menon… - arXiv preprint arXiv …, 2023 - arxiv.org
We introduce a video framework for modeling the association between verbal and non-
verbal communication during dyadic conversation. Given the input speech of a speaker, our …

Investigating how speech and animation realism influence the perceived personality of virtual characters and agents

S Thomas, Y Ferstl, R McDonnell… - 2022 IEEE Conference …, 2022 - ieeexplore.ieee.org
The portrayed personality of virtual characters and agents is understood to influence how we
perceive and engage with digital applications. Understanding how the features of speech …

S3: Speech, Script and Scene driven Head and Eye Animation

Y Pan, R Agrawal, K Singh - ACM Transactions on Graphics (TOG), 2024 - dl.acm.org
We present S 3, a novel approach to generating expressive, animator-centric 3D head and
eye animation of characters in conversation. Given speech audio, a Directorial script and a …