Motion-i2v: Consistent and controllable image-to-video generation with explicit motion modeling

X Shi, Z Huang, FY Wang, W Bian, D Li… - ACM SIGGRAPH 2024 …, 2024 - dl.acm.org
We introduce Motion-I2V, a novel framework for consistent and controllable text-guided
image-to-video generation (I2V). In contrast to previous methods that directly learn the …

Diffusion model-based video editing: A survey

W Sun, RC Tu, J Liao, D Tao - arxiv preprint arxiv:2407.07111, 2024 - arxiv.org
The rapid development of diffusion models (DMs) has significantly advanced image and
video applications, making" what you want is what you see" a reality. Among these, video …

Tc4d: Trajectory-conditioned text-to-4d generation

S Bahmani, X Liu, W Yifan, I Skorokhodov… - … on Computer Vision, 2024 - Springer
Recent techniques for text-to-4D generation synthesize dynamic 3D scenes using
supervision from pre-trained text-to-video models. However, existing representations, such …

Dragvideo: Interactive drag-style video editing

Y Deng, R Wang, Y Zhang, YW Tai, CK Tang - European Conference on …, 2024 - Springer
Video generation models have shown their superior ability to generate photo-realistic video.
However, how to accurately control (or edit) the video remains a formidable challenge. The …

AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing Tasks

M Ku, C Wei, W Ren, H Yang, W Chen - Transactions on Machine …, 2024 - openreview.net
In the dynamic field of digital content creation using generative models, state-of-the-art video
editing models still do not offer the level of quality and control that users desire. Previous …

Efficient diffusion models: A comprehensive survey from principles to practices

Z Ma, Y Zhang, G Jia, L Zhao, Y Ma, M Ma… - arxiv preprint arxiv …, 2024 - arxiv.org
As one of the most popular and sought-after generative models in the recent years, diffusion
models have sparked the interests of many researchers and steadily shown excellent …

Anyv2v: A plug-and-play framework for any video-to-video editing tasks

M Ku, C Wei, W Ren, H Yang, W Chen - arxiv preprint arxiv:2403.14468, 2024 - arxiv.org
Video-to-video editing involves editing a source video along with additional control (such as
text prompts, subjects, or styles) to generate a new video that aligns with the source video …

Compositional 3d-aware video generation with llm director

H Zhu, T He, A Tang, J Guo, Z Chen, J Bian - arxiv preprint arxiv …, 2024 - arxiv.org
Significant progress has been made in text-to-video generation through the use of powerful
generative models and large-scale internet data. However, substantial challenges remain in …

SAVE: Protagonist Diversification with S tructure A gnostic V ideo E diting

Y Song, W Shin, J Lee, J Kim, N Kwak - European Conference on …, 2024 - Springer
Driven by the upsurge progress in text-to-image (T2I) generation models, text-to-video (T2V)
generation has experienced a significant advance as well. Accordingly, tasks such as …

Spectral motion alignment for video motion transfer using diffusion models

GY Park, H Jeong, SW Lee, JC Ye - arxiv preprint arxiv:2403.15249, 2024 - arxiv.org
The evolution of diffusion models has greatly impacted video generation and understanding.
Particularly, text-to-video diffusion models (VDMs) have significantly facilitated the …