Learning to listen: Modeling non-deterministic dyadic facial motion
We present a framework for modeling interactional communication in dyadic conversations:
given multimodal inputs of a speaker, we autoregressively output multiple possibilities of …
given multimodal inputs of a speaker, we autoregressively output multiple possibilities of …
Can language models learn to listen?
E Ng, S Subramanian, D Klein… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present a framework for generating appropriate facial responses from a listener in
dyadic social interactions based on the speaker's words. Given an input transcription of the …
dyadic social interactions based on the speaker's words. Given an input transcription of the …
ZeroEGGS: Zero‐shot Example‐based Gesture Generation from Speech
We present ZeroEGGS, a neural network framework for speech‐driven gesture generation
with zero‐shot style control by example. This means style can be controlled via only a short …
with zero‐shot style control by example. This means style can be controlled via only a short …
What affects the usage of artificial conversational agents? An agent personality and love theory perspective
Artificial-intelligence (AI) powered conversational (CAI) agents have been growing in
popularity. Like human personality, CAI agent personality can also impact the relationship …
popularity. Like human personality, CAI agent personality can also impact the relationship …
Robotic mental well-being coaches for the workplace: An in-the-wild study on form
The World Health Organization recommends that employers take action to protect and
promote mental well-being at work. However, the extent to which these recommended …
promote mental well-being at work. However, the extent to which these recommended …
ExpressGesture: Expressive gesture generation from speech through database matching
Co‐speech gestures are a vital ingredient in making virtual agents more human‐like and
engaging. Automatically generated gestures based on speech‐input often lack realistic and …
engaging. Automatically generated gestures based on speech‐input often lack realistic and …
The MuSe 2024 multimodal sentiment analysis challenge: Social perception and humor recognition
The Multimodal Sentiment Analysis Challenge (MuSe) 2024 addresses two contemporary
multimodal affect and sentiment analysis problems: In the Social Perception Sub-Challenge …
multimodal affect and sentiment analysis problems: In the Social Perception Sub-Challenge …
Affective faces for goal-driven dyadic communication
We introduce a video framework for modeling the association between verbal and non-
verbal communication during dyadic conversation. Given the input speech of a speaker, our …
verbal communication during dyadic conversation. Given the input speech of a speaker, our …
Investigating how speech and animation realism influence the perceived personality of virtual characters and agents
The portrayed personality of virtual characters and agents is understood to influence how we
perceive and engage with digital applications. Understanding how the features of speech …
perceive and engage with digital applications. Understanding how the features of speech …
S3: Speech, Script and Scene driven Head and Eye Animation
We present S 3, a novel approach to generating expressive, animator-centric 3D head and
eye animation of characters in conversation. Given speech audio, a Directorial script and a …
eye animation of characters in conversation. Given speech audio, a Directorial script and a …