A survey on lora of large language models

Y Mao, Y Ge, Y Fan, W Xu, Y Mi, Z Hu… - Frontiers of Computer …, 2025 - Springer
Abstract Low-Rank Adaptation (LoRA), which updates the dense neural network layers with
pluggable low-rank matrices, is one of the best performed parameter efficient fine-tuning …

Motiondirector: Motion customization of text-to-video diffusion models

R Zhao, Y Gu, JZ Wu, DJ Zhang, JW Liu, W Wu… - … on Computer Vision, 2024 - Springer
Large-scale pre-trained diffusion models have exhibited remarkable capabilities in diverse
video generations. Given a set of video clips of the same motion concept, the task of Motion …

Customizing text-to-image models with a single image pair

M Jones, SY Wang, N Kumari, D Bau… - SIGGRAPH Asia 2024 …, 2024 - dl.acm.org
Art reinterpretation is the practice of creating a variation of a reference work, making a paired
artwork that exhibits a distinct artistic style. We ask if such an image pair can be used to …

Videoshop: Localized semantic video editing with noise-extrapolated diffusion inversion

X Fan, A Bhattad, R Krishna - European Conference on Computer Vision, 2024 - Springer
We introduce Videoshop, a training-free video editing algorithm for localized semantic edits.
Videoshop allows users to use any editing software, including Photoshop and generative …

Spectral motion alignment for video motion transfer using diffusion models

GY Park, H Jeong, SW Lee, JC Ye - arxiv preprint arxiv:2403.15249, 2024 - arxiv.org
The evolution of diffusion models has greatly impacted video generation and understanding.
Particularly, text-to-video diffusion models (VDMs) have significantly facilitated the …

Motrans: Customized motion transfer with text-driven video diffusion models

X Li, X Jia, Q Wang, H Diao, M Ge, P Li, Y He… - Proceedings of the 32nd …, 2024 - dl.acm.org
Existing pretrained text-to-video (T2V) models have demonstrated impressive abilities in
generating realistic videos with basic motion or camera movement. However, these models …

OmniCreator: Self-Supervised Unified Generation with Universal Editing

H Chen, L Wang, H Yang, SN Lim - arxiv preprint arxiv:2412.02114, 2024 - arxiv.org
We introduce OmniCreator, a novel framework that can conduct text-prompted unified
(image+ video) generation as well as editing all in one place. OmniCreator acquires …

Human motion video generation: A survey

H Xue, X Luo, Z Hu, X Zhang, X **ang, Y Dai, J Liu… - Authorea …, 2024 - techrxiv.org
Human motion video generation has garnered significant research interest due to its broad
applications, enabling innovations such as photorealistic singing heads or dynamic avatars …

A Survey on Personalized Content Synthesis with Diffusion Models

X Zhang, XY Wei, W Zhang, J Wu, Z Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent advancements in generative models have significantly impacted content creation,
leading to the emergence of Personalized Content Synthesis (PCS). With a small set of user …

VFX Creator: Animated Visual Effect Generation with Controllable Diffusion Transformer

X Liu, A Zeng, W Xue, H Yang, W Luo, Q Liu… - arxiv preprint arxiv …, 2025 - arxiv.org
Crafting magic and illusions is one of the most thrilling aspects of filmmaking, with visual
effects (VFX) serving as the powerhouse behind unforgettable cinematic experiences. While …