Is sora a world simulator? a comprehensive survey on general world models and beyond

Z Zhu, X Wang, W Zhao, C Min, N Deng, M Dou… - arxiv preprint arxiv …, 2024 - arxiv.org
General world models represent a crucial pathway toward achieving Artificial General
Intelligence (AGI), serving as the cornerstone for various applications ranging from virtual …

Learning-to-cache: Accelerating diffusion transformer via layer caching

X Ma, G Fang, M Bi Mi, X Wang - Advances in Neural …, 2025 - proceedings.neurips.cc
Diffusion Transformers have recently demonstrated unprecedented generative capabilities
for various tasks. The encouraging results, however, come with the cost of slow inference …

Efficient diffusion models: A comprehensive survey from principles to practices

Z Ma, Y Zhang, G Jia, L Zhao, Y Ma, M Ma… - arxiv preprint arxiv …, 2024 - arxiv.org
As one of the most popular and sought-after generative models in the recent years, diffusion
models have sparked the interests of many researchers and steadily shown excellent …

Mixture of efficient diffusion experts through automatic interval and sub-network selection

A Ganjdanesh, Y Kang, Y Liu, R Zhang, Z Lin… - … on Computer Vision, 2024 - Springer
Diffusion probabilistic models can generate high-quality samples. Yet, their sampling
process requires numerous denoising steps, making it slow and computationally intensive …

Dynamic diffusion transformer

W Zhao, Y Han, J Tang, K Wang, Y Song… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion Transformer (DiT), an emerging diffusion model for image generation, has
demonstrated superior performance but suffers from substantial computational costs. Our …

Remix-DiT: Mixing Diffusion Transformers for Multi-Expert Denoising

G Fang, X Ma, X Wang - Advances in Neural Information …, 2025 - proceedings.neurips.cc
Transformer-based diffusion models have achieved significant advancements across a
variety of generative tasks. However, producing high-quality outputs typically necessitates …

HarmoniCa: Harmonizing Training and Inference for Better Feature Cache in Diffusion Transformer Acceleration

Y Huang, Z Wang, R Gong, J Liu, X Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion Transformers (DiTs) have gained prominence for outstanding scalability and
extraordinary performance in generative tasks. However, their considerable inference costs …

Recurrent Diffusion for Large-Scale Parameter Generation

K Wang, D Tang, W Zhao, Y You - arxiv preprint arxiv:2501.11587, 2025 - arxiv.org
Parameter generation has struggled to scale up for a long time, significantly limiting its range
of applications. In this study, we introduce\textbf {R} ecurrent diffusion for large-scale\textbf …

Denoising Task Difficulty-based Curriculum for Training Diffusion Models

JY Kim, H Go, S Kwon, HG Kim - arxiv preprint arxiv:2403.10348, 2024 - arxiv.org
Diffusion-based generative models have emerged as powerful tools in the realm of
generative modeling. Despite extensive research on denoising across various timesteps …

Diffusion Model Compression for Image-to-Image Translation

G Kim, B Kim, E Park, S Cho - Proceedings of the Asian …, 2024 - openaccess.thecvf.com
As recent advances in large-scale Text-to-Image (T2I) diffusion models have yielded
remarkable high-quality image generation, diverse downstream Image-to-Image (I2I) …