Revideo: Remake a video with motion and content control

C Mou, M Cao, X Wang, Z Zhang… - Advances in Neural …, 2025 - proceedings.neurips.cc
Despite significant advancements in video generation and editing using diffusion models,
achieving accurate and localized video editing remains a substantial challenge …

Diffusion model-based video editing: A survey

W Sun, RC Tu, J Liao, D Tao - arxiv preprint arxiv:2407.07111, 2024 - arxiv.org
The rapid development of diffusion models (DMs) has significantly advanced image and
video applications, making" what you want is what you see" a reality. Among these, video …

Llms meet multimodal generation and editing: A survey

Y He, Z Liu, J Chen, Z Tian, H Liu, X Chi, R Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
With the recent advancement in large language models (LLMs), there is a growing interest in
combining LLMs with multimodal learning. Previous surveys of multimodal large language …

Video editing via factorized diffusion distillation

U Singer, A Zohar, Y Kirstain, S Sheynin… - … on Computer Vision, 2024 - Springer
Abstract We introduce Emu Video Edit (EVE), a model that establishes a new state-of-the art
in video editing without relying on any supervised video editing data. To develop EVE we …

Rethinking The Training And Evaluation of Rich-Context Layout-to-Image Generation

J Cheng, Z Zhao, T He, T **ao… - Advances in Neural …, 2025 - proceedings.neurips.cc
Recent advancements in generative models have significantly enhanced their capacity for
image generation, enabling a wide range of applications such as image editing, completion …

AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing Tasks

M Ku, C Wei, W Ren, H Yang, W Chen - arxiv preprint arxiv:2403.14468, 2024 - arxiv.org
In the dynamic field of digital content creation using generative models, state-of-the-art video
editing models still do not offer the level of quality and control that users desire. Previous …

NaRCan: Natural Refined Canonical Image with Integration of Diffusion Prior for Video Editing

TH Chen, JW Chan, HS Shiu, SH Yen… - Advances in Neural …, 2025 - proceedings.neurips.cc
We propose a video editing framework, NaRCan, which integrates a hybrid deformation field
and diffusion prior to generate high-quality natural canonical images to represent the input …

OmniCreator: Self-Supervised Unified Generation with Universal Editing

H Chen, L Wang, H Yang, SN Lim - arxiv preprint arxiv:2412.02114, 2024 - arxiv.org
We introduce OmniCreator, a novel framework that can conduct text-prompted unified
(image+ video) generation as well as editing all in one place. OmniCreator acquires …

Generative Video Propagation

S Liu, T Wang, JH Wang, Q Liu, Z Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large-scale video generation models have the inherent ability to realistically model natural
scenes. In this paper, we demonstrate that through a careful design of a generative video …

RelightVid: Temporal-Consistent Diffusion Model for Video Relighting

Y Fang, Z Sun, S Zhang, T Wu, Y Xu, P Zhang… - arxiv preprint arxiv …, 2025 - arxiv.org
Diffusion models have demonstrated remarkable success in image generation and editing,
with recent advancements enabling albedo-preserving image relighting. However, applying …