A survey of multimodal-guided image editing with text-to-image diffusion models

X Shuai, H Ding, X Ma, R Tu, YG Jiang… - arxiv preprint arxiv …, 2024 - arxiv.org
Image editing aims to edit the given synthetic or real image to meet the specific requirements
from users. It is widely studied in recent years as a promising and challenging field of …

Advances in diffusion models for image data augmentation: A review of methods, models, evaluation metrics and future research directions

P Alimisis, I Mademlis, P Radoglou-Grammatikis… - Artificial Intelligence …, 2025 - Springer
Image data augmentation constitutes a critical methodology in modern computer vision
tasks, since it can facilitate towards enhancing the diversity and quality of training datasets; …

Diffusion model-based image editing: A survey

Y Huang, J Huang, Y Liu, M Yan, J Lv, J Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
Denoising diffusion models have emerged as a powerful tool for various image generation
and editing tasks, facilitating the synthesis of visual content in an unconditional or input …

Inversion-free image editing with natural language

S Xu, Y Huang, J Pan, Z Ma, J Chai - arxiv preprint arxiv:2312.04965, 2023 - arxiv.org
Despite recent advances in inversion-based editing, text-guided image manipulation
remains challenging for diffusion models. The primary bottlenecks include 1) the time …

Scedit: Efficient and controllable image diffusion generation via skip connection editing

Z Jiang, C Mao, Y Pan, Z Han… - Proceedings of the …, 2024 - openaccess.thecvf.com
Image diffusion models have been utilized in various tasks such as text-to-image generation
and controllable image synthesis. Recent research has introduced tuning methods that …

Inversion-free image editing with language-guided diffusion models

S Xu, Y Huang, J Pan, Z Ma… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Despite recent advances in inversion-based editing text-guided image manipulation
remains challenging for diffusion models. The primary bottlenecks include 1) the time …

Smoodi: Stylized motion diffusion model

L Zhong, Y **e, V Jampani, D Sun, H Jiang - European Conference on …, 2024 - Springer
We introduce a novel Stylized Motion Diffusion model, dubbed SMooDi, to generate stylized
motion driven by content texts and style motion sequences. Unlike existing methods that …

Efficient diffusion models: A comprehensive survey from principles to practices

Z Ma, Y Zhang, G Jia, L Zhao, Y Ma, M Ma… - arxiv preprint arxiv …, 2024 - arxiv.org
As one of the most popular and sought-after generative models in the recent years, diffusion
models have sparked the interests of many researchers and steadily shown excellent …

Replication in visual diffusion models: A survey and outlook

W Wang, Y Sun, Z Yang, Z Hu, Z Tan… - arxiv preprint arxiv …, 2024 - arxiv.org
Visual diffusion models have revolutionized the field of creative AI, producing high-quality
and diverse content. However, they inevitably memorize training images or videos …

Unictrl: Improving the spatiotemporal consistency of text-to-video diffusion models via training-free unified attention control

T **a, X Chen, S Xu - arxiv preprint arxiv:2403.02332, 2024 - arxiv.org
Video Diffusion Models have been developed for video generation, usually integrating text
and image conditioning to enhance control over the generated content. Despite the …