Sadtalker: Learning realistic 3d motion coefficients for stylized audio-driven single image talking face animation

W Zhang, X Cun, X Wang, Y Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Generating talking head videos through a face image and a piece of speech audio still
contains many challenges. ie, unnatural head movement, distorted expression, and identity …

Next3d: Generative neural texture rasterization for 3d-aware head avatars

J Sun, X Wang, L Wang, X Li… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract 3D-aware generative adversarial networks (GANs) synthesize high-fidelity and
multi-view-consistent facial images using only collections of single-view 2D imagery …

Self-supervised learning of adversarial example: Towards good generalizations for deepfake detection

L Chen, Y Zhang, Y Song, L Liu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Recent studies in deepfake detection have yielded promising results when the training and
testing face forgeries are from the same dataset. However, the problem remains challenging …

[HTML][HTML] Talking human face generation: A survey

M Toshpulatov, W Lee, S Lee - Expert Systems with Applications, 2023 - Elsevier
Talking human face generation aims at synthesizing a natural human face that talks in
correspondence to the given text or audio series. Implementing the recently developed …

Diffused heads: Diffusion models beat gans on talking-face generation

M Stypułkowski, K Vougioukas, S He… - Proceedings of the …, 2024 - openaccess.thecvf.com
Talking face generation has historically struggled to produce head movements and natural
facial expressions without guidance from additional reference videos. Recent developments …

Otavatar: One-shot talking face avatar with controllable tri-plane rendering

Z Ma, X Zhu, GJ Qi, Z Lei… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Controllability, generalizability and efficiency are the major objectives of constructing face
avatars represented by neural implicit field. However, existing methods have not managed …

Progressive disentangled representation learning for fine-grained controllable talking head synthesis

D Wang, Y Deng, Z Yin, HY Shum… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present a novel one-shot talking head synthesis method that achieves disentangled and
fine-grained control over lip motion, eye gaze&blink, head pose, and emotional expression …

Stylesync: High-fidelity generalized and personalized lip sync in style-based generator

J Guan, Z Zhang, H Zhou, T Hu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Despite recent advances in syncing lip movements with any audio waves, current methods
still struggle to balance generation quality and the model's generalization ability. Previous …

Metaportrait: Identity-preserving talking head generation with fast personalized adaptation

B Zhang, C Qi, P Zhang, B Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this work, we propose an ID-preserving talking head generation framework, which
advances previous methods in two aspects. First, as opposed to interpolating from sparse …

Styletalk: One-shot talking head generation with controllable speaking styles

Y Ma, S Wang, Z Hu, C Fan, T Lv, Y Ding… - Proceedings of the …, 2023 - ojs.aaai.org
Different people speak with diverse personalized speaking styles. Although existing one-
shot talking head methods have made significant progress in lip sync, natural facial …