Listen, denoise, action! audio-driven motion synthesis with diffusion models

S Alexanderson, R Nagy, J Beskow… - ACM Transactions on …, 2023 - dl.acm.org
Diffusion models have experienced a surge of interest as highly expressive yet efficiently
trainable probabilistic models. We show that these models are an excellent fit for …

Human motion diffusion as a generative prior

Y Shafir, G Tevet, R Kapon, AH Bermano - arXiv preprint arXiv:2303.01418, 2023 - arxiv.org
Recent work has demonstrated the significant potential of denoising diffusion models for
generating human motion, including text-to-motion capabilities. However, these methods are …

Clipface: Text-guided editing of textured 3d morphable models

S Aneja, J Thies, A Dai, M Nießner - ACM SIGGRAPH 2023 Conference …, 2023 - dl.acm.org
We propose ClipFace, a novel self-supervised approach for text-guided editing of textured
3D morphable model of faces. Specifically, we employ user-friendly language prompts to …

Single motion diffusion

S Raab, I Leibovitch, G Tevet, M Arar… - arXiv preprint arXiv …, 2023 - arxiv.org
Synthesizing realistic animations of humans, animals, and even imaginary creatures, has
long been a goal for artists and computer graphics professionals. Compared to the imaging …

Moconvq: Unified physics-based motion control via scalable discrete representations

H Yao, Z Song, Y Zhou, T Ao, B Chen… - ACM Transactions on …, 2024 - dl.acm.org
In this work, we present MoConVQ, a novel unified framework for physics-based motion
control leveraging scalable discrete representations. Building upon vector quantized …

MotionFix: Text-driven 3d human motion editing

N Athanasiou, A Cseke, M Diomataris… - SIGGRAPH Asia 2024 …, 2024 - dl.acm.org
The focus of this paper is 3D motion editing. Given a 3D human motion and a textual
description of the desired modification, our goal is to generate an edited motion as …

Synthesizing long-term human motions with diffusion models via coherent sampling

Z Yang, B Su, JR Wen - Proceedings of the 31st ACM International …, 2023 - dl.acm.org
Text-to-motion generation has gained increasing attention, but most existing methods are
limited to generating short-term motions that correspond to a single sentence describing a …

Monkey see, monkey do: Harnessing self-attention in motion diffusion for zero-shot motion transfer

S Raab, I Gat, N Sala, G Tevet… - SIGGRAPH Asia 2024 …, 2024 - dl.acm.org
Given the remarkable results of motion synthesis with diffusion models, a natural question
arises: how can we effectively leverage these models for motion editing? Existing diffusion …

[PDF][PDF] Ai-generated content (aigc) for various data modalities: A survey

LG Foo, H Rahmani, J Liu - arXiv preprint arXiv:2308.14177, 2023 - researchgate.net
Amidst the rapid advancement of artificial intelligence (AI), the development of content
generation techniques stands out as one of the most captivating and widely discussed topics …

SATO: Stable Text-to-Motion Framework

W Chen, H Xiao, E Zhang, L Hu, L Wang… - Proceedings of the …, 2024 - dl.acm.org
Is the Text to Motion model robust? Recent advancements in Text to Motion models primarily
stem from more accurate predictions of specific actions. However, the text modality typically …