Efficient diffusion models: A comprehensive survey from principles to practices

Z Ma, Y Zhang, G Jia, L Zhao, Y Ma, M Ma… - arxiv preprint arxiv …, 2024 - arxiv.org
As one of the most popular and sought-after generative models in the recent years, diffusion
models have sparked the interests of many researchers and steadily shown excellent …

Migc++: Advanced multi-instance generation controller for image synthesis

D Zhou, Y Li, F Ma, Z Yang… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
We introduce the Multi-Instance Generation (MIG) task, which focuses on generating
multiple instances within a single image, each accurately placed at predefined positions with …

Stablegarment: Garment-centric generation via stable diffusion

R Wang, H Guo, J Liu, H Li, H Zhao, X Tang… - arxiv preprint arxiv …, 2024 - arxiv.org
In this paper, we introduce StableGarment, a unified framework to tackle garment-centric
(GC) generation tasks, including GC text-to-image, controllable GC text-to-image, stylized …

Ms-diffusion: Multi-subject zero-shot image personalization with layout guidance

X Wang, S Fu, Q Huang, W He, H Jiang - arxiv preprint arxiv:2406.07209, 2024 - arxiv.org
Recent advancements in text-to-image generation models have dramatically enhanced the
generation of photorealistic images from textual prompts, leading to an increased interest in …

-ECLIPSE: Multi-Concept Personalized Text-to-Image Diffusion Models by Leveraging CLIP Latent Space

M Patel, S Jung, C Baral, Y Yang - arxiv preprint arxiv:2402.05195, 2024 - arxiv.org
Despite the recent advances in personalized text-to-image (P-T2I) generative models, it
remains challenging to perform finetuning-free multi-subject-driven T2I in a resource …

Editworld: Simulating world dynamics for instruction-following image editing

L Yang, B Zeng, J Liu, H Li, M Xu, W Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion models have significantly improved the performance of image editing. Existing
methods realize various approaches to achieve high-quality image editing, including but not …

Freetuner: Any subject in any style with training-free diffusion

Y Xu, Z Wang, J **ao, W Liu, L Chen - arxiv preprint arxiv:2405.14201, 2024 - arxiv.org
With the advance of diffusion models, various personalized image generation methods have
been proposed. However, almost all existing work only focuses on either subject-driven or …

Foodfusion: A novel approach for food image composition via diffusion models

C Shi, X Wang, S Shi, X Wang, M Zhu, N Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
Food image composition requires the use of existing dish images and background images to
synthesize a natural new image, while diffusion models have made significant …

3dis: Depth-driven decoupled instance synthesis for text-to-image generation

D Zhou, J **e, Z Yang, Y Yang - arxiv preprint arxiv:2410.12669, 2024 - arxiv.org
The increasing demand for controllable outputs in text-to-image generation has spurred
advancements in multi-instance generation (MIG), allowing users to define both instance …

Masterweaver: Taming editability and face identity for personalized text-to-image generation

Y Wei, Z Ji, J Bai, H Zhang, L Zhang, W Zuo - European Conference on …, 2024 - Springer
Abstract Text-to-image (T2I) diffusion models have shown significant success in
personalized text-to-image generation, which aims to generate novel images with human …