Vasa-1: Lifelike audio-driven talking faces generated in real time

S Xu, G Chen, YX Guo, J Yang, C Li… - Advances in …, 2025 - proceedings.neurips.cc
We introduce VASA, a framework for generating lifelike talking faces with appealing visual
affective skills (VAS) given a single static image and a speech audio clip. Our premiere …

Facechain-imagineid: Freely crafting high-fidelity diverse talking faces from disentangled audio

C Xu, Y Liu, J **ng, W Wang, M Sun… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this paper we abstract the process of people hearing speech extracting meaningful cues
and creating various dynamically audio-consistent talking faces termed Listening and …

Deepfake generation and detection: A benchmark and survey

G Pei, J Zhang, M Hu, Z Zhang, C Wang, Y Wu… - arxiv preprint arxiv …, 2024 - arxiv.org
Deepfake is a technology dedicated to creating highly realistic facial images and videos
under specific conditions, which has significant application potential in fields such as …

Portrait4d: Learning one-shot 4d head avatar synthesis using synthetic data

Y Deng, D Wang, X Ren, X Chen… - Proceedings of the …, 2024 - openaccess.thecvf.com
Existing one-shot 4D head synthesis methods usually learn from monocular videos with the
aid of 3DMM reconstruction yet the latter is evenly challenging which restricts them from …

Vlogger: Multimodal diffusion for embodied avatar synthesis

E Corona, A Zanfir, EG Bazavan, N Kolotouros… - arxiv preprint arxiv …, 2024 - arxiv.org
We propose VLOGGER, a method for audio-driven human video generation from a single
input image of a person, which builds on the success of recent generative diffusion models …

Hallo: Hierarchical audio-driven visual synthesis for portrait image animation

M Xu, H Li, Q Su, H Shang, L Zhang, C Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
The field of portrait image animation, driven by speech audio input, has experienced
significant advancements in the generation of realistic and dynamic portraits. This research …

Multilingual video dubbing—a technology review and current challenges

D Bigioi, P Corcoran - Frontiers in signal processing, 2023 - frontiersin.org
The proliferation of multi-lingual content on today's streaming services has created a need
for automated multi-lingual dubbing tools. In this article, current state-of-the-art approaches …

A Comprehensive Survey on Human Video Generation: Challenges, Methods, and Insights

W Lei, J Wang, F Ma, G Huang, L Liu - arxiv preprint arxiv:2407.08428, 2024 - arxiv.org
Human video generation is a dynamic and rapidly evolving task that aims to synthesize 2D
human body video sequences with generative models given control conditions such as text …

Learning to generate conditional tri-plane for 3d-aware expression controllable portrait animation

T Ki, D Min, G Chae - European Conference on Computer Vision, 2024 - Springer
In this paper, we present Export3D, a one-shot 3D-aware portrait animation method that is
able to control the facial expression and camera view of a given portrait image. To achieve …

VOODOO 3D: volumetric portrait disentanglement for one-shot 3d head reenactment

P Tran, E Zakharov, LN Ho, AT Tran… - Proceedings of the …, 2024 - openaccess.thecvf.com
We present a 3D-aware one-shot head reenactment method based on a fully volumetric
neural disentanglement framework for source appearance and driver expressions. Our …