A survey on video diffusion models

Z **ng, Q Feng, H Chen, Q Dai, H Hu, H Xu… - ACM Computing …, 2024 - dl.acm.org
The recent wave of AI-generated content (AIGC) has witnessed substantial success in
computer vision, with the diffusion model playing a crucial role in this achievement. Due to …

Enhancing deep reinforcement learning: A tutorial on generative diffusion models in network optimization

H Du, R Zhang, Y Liu, J Wang, Y Lin… - … Surveys & Tutorials, 2024 - ieeexplore.ieee.org
Generative Diffusion Models (GDMs) have emerged as a transformative force in the realm of
Generative Artificial Intelligence (GenAI), demonstrating their versatility and efficacy across …

Animate anyone: Consistent and controllable image-to-video synthesis for character animation

L Hu - Proceedings of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Character Animation aims to generating character videos from still images through driving
signals. Currently diffusion models have become the mainstream in visual generation …

Vbench: Comprehensive benchmark suite for video generative models

Z Huang, Y He, J Yu, F Zhang, C Si… - Proceedings of the …, 2024 - openaccess.thecvf.com
Video generation has witnessed significant advancements yet evaluating these models
remains a challenge. A comprehensive evaluation benchmark for video generation is …

Videocomposer: Compositional video synthesis with motion controllability

X Wang, H Yuan, S Zhang, D Chen… - Advances in …, 2023 - proceedings.neurips.cc
The pursuit of controllability as a higher standard of visual content creation has yielded
remarkable progress in customizable image synthesis. However, achieving controllable …

Dynamicrafter: Animating open-domain images with video diffusion priors

J **ng, M **a, Y Zhang, H Chen, W Yu, H Liu… - … on Computer Vision, 2024 - Springer
Animating a still image offers an engaging visual experience. Traditional image animation
techniques mainly focus on animating natural scenes with stochastic dynamics (eg clouds …

Rich human feedback for text-to-image generation

Y Liang, J He, G Li, P Li, A Klimovskiy… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Recent Text-to-Image (T2I) generation models such as Stable Diffusion and Imagen
have made significant progress in generating high-resolution images based on text …

Storydiffusion: Consistent self-attention for long-range image and video generation

Y Zhou, D Zhou, MM Cheng… - Advances in Neural …, 2025 - proceedings.neurips.cc
For recent diffusion-based generative models, maintaining consistent content across a
series of generated images, especially those containing subjects and complex details …

Motiondirector: Motion customization of text-to-video diffusion models

R Zhao, Y Gu, JZ Wu, DJ Zhang, JW Liu, W Wu… - … on Computer Vision, 2024 - Springer
Large-scale pre-trained diffusion models have exhibited remarkable capabilities in diverse
video generations. Given a set of video clips of the same motion concept, the task of Motion …

Disco: Disentangled control for realistic human dance generation

T Wang, L Li, K Lin, Y Zhai, CC Lin… - Proceedings of the …, 2024 - openaccess.thecvf.com
Generative AI has made significant strides in computer vision particularly in text-driven
image/video synthesis (T2I/T2V). Despite the notable advancements it remains challenging …