A survey on video diffusion models

Z **ng, Q Feng, H Chen, Q Dai, H Hu, H Xu… - ACM Computing …, 2024 - dl.acm.org
The recent wave of AI-generated content (AIGC) has witnessed substantial success in
computer vision, with the diffusion model playing a crucial role in this achievement. Due to …

Image synthesis with adversarial networks: A comprehensive survey and case studies

P Shamsolmoali, M Zareapoor, E Granger, H Zhou… - Information …, 2021 - Elsevier
Abstract Generative Adversarial Networks (GANs) have been extremely successful in
various application domains such as computer vision, medicine, and natural language …

Dynamicrafter: Animating open-domain images with video diffusion priors

J **ng, M **a, Y Zhang, H Chen, W Yu, H Liu… - … on Computer Vision, 2024 - Springer
Animating a still image offers an engaging visual experience. Traditional image animation
techniques mainly focus on animating natural scenes with stochastic dynamics (eg clouds …

Latte: Latent diffusion transformer for video generation

X Ma, Y Wang, G Jia, X Chen, Z Liu, YF Li… - arxiv preprint arxiv …, 2024 - arxiv.org
We propose a novel Latent Diffusion Transformer, namely Latte, for video generation. Latte
first extracts spatio-temporal tokens from input videos and then adopts a series of …

Latent video diffusion models for high-fidelity long video generation

Y He, T Yang, Y Zhang, Y Shan, Q Chen - arxiv preprint arxiv:2211.13221, 2022 - arxiv.org
AI-generated content has attracted lots of attention recently, but photo-realistic video
synthesis is still challenging. Although many attempts using GANs and autoregressive …

Video probabilistic diffusion models in projected latent space

S Yu, K Sohn, S Kim, J Shin - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Despite the remarkable progress in deep generative models, synthesizing high-resolution
and temporally coherent videos still remains a challenge due to their high-dimensionality …

Videofusion: Decomposed diffusion models for high-quality video generation

Z Luo, D Chen, Y Zhang, Y Huang, L Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
A diffusion probabilistic model (DPM), which constructs a forward diffusion process by
gradually adding noise to data points and learns the reverse denoising process to generate …

Conditional image-to-video generation with latent flow diffusion models

H Ni, C Shi, K Li, SX Huang… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Conditional image-to-video (cI2V) generation aims to synthesize a new plausible video
starting from an image (eg, a person's face) and a condition (eg, an action class label like …

Simda: Simple diffusion adapter for efficient video generation

Z **ng, Q Dai, H Hu, Z Wu… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
The recent wave of AI-generated content has witnessed the great development and success
of Text-to-Image (T2I) technologies. By contrast Text-to-Video (T2V) still falls short of …

Stylegan-v: A continuous video generator with the price, image quality and perks of stylegan2

I Skorokhodov, S Tulyakov… - Proceedings of the …, 2022 - openaccess.thecvf.com
Videos show continuous events, yet most--if not all--video synthesis frameworks treat them
discretely in time. In this work, we think of videos of what they should be--time-continuous …