A survey on video diffusion models

Z **ng, Q Feng, H Chen, Q Dai, H Hu, H Xu… - ACM Computing …, 2024 - dl.acm.org
The recent wave of AI-generated content (AIGC) has witnessed substantial success in
computer vision, with the diffusion model playing a crucial role in this achievement. Due to …

AI-generated characters for supporting personalized learning and well-being

P Pataranutaporn, V Danry, J Leong… - Nature Machine …, 2021 - nature.com
Advancements in machine learning have recently enabled the hyper-realistic synthesis of
prose, images, audio and video data, in what is referred to as artificial intelligence (AI) …

Dynamicrafter: Animating open-domain images with video diffusion priors

J **ng, M **a, Y Zhang, H Chen, W Yu, H Liu… - … on Computer Vision, 2024 - Springer
Animating a still image offers an engaging visual experience. Traditional image animation
techniques mainly focus on animating natural scenes with stochastic dynamics (eg clouds …

Videofusion: Decomposed diffusion models for high-quality video generation

Z Luo, D Chen, Y Zhang, Y Huang, L Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
A diffusion probabilistic model (DPM), which constructs a forward diffusion process by
gradually adding noise to data points and learns the reverse denoising process to generate …

Conditional image-to-video generation with latent flow diffusion models

H Ni, C Shi, K Li, SX Huang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Conditional image-to-video (cI2V) generation aims to synthesize a new plausible video
starting from an image (eg, a person's face) and a condition (eg, an action class label like …

Sadtalker: Learning realistic 3d motion coefficients for stylized audio-driven single image talking face animation

W Zhang, X Cun, X Wang, Y Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Generating talking head videos through a face image and a piece of speech audio still
contains many challenges. ie, unnatural head movement, distorted expression, and identity …

Champ: Controllable and consistent human image animation with 3d parametric guidance

S Zhu, JL Chen, Z Dai, Z Dong, Y Xu, X Cao… - … on Computer Vision, 2024 - Springer
In this study, we introduce a methodology for human image animation by leveraging a 3D
human parametric model within a latent diffusion framework to enhance shape alignment …

Animate anyone: Consistent and controllable image-to-video synthesis for character animation

L Hu - Proceedings of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Character Animation aims to generating character videos from still images through driving
signals. Currently diffusion models have become the mainstream in visual generation …

Avatarclip: Zero-shot text-driven generation and animation of 3d avatars

F Hong, M Zhang, L Pan, Z Cai, L Yang… - arxiv preprint arxiv …, 2022 - arxiv.org
3D avatar creation plays a crucial role in the digital age. However, the whole production
process is prohibitively time-consuming and labor-intensive. To democratize this technology …

Generative image dynamics

Z Li, R Tucker, N Snavely… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
We present an approach to modeling an image-space prior on scene motion. Our prior is
learned from a collection of motion trajectories extracted from real video sequences …