Evolutionary computation in the era of large language model: Survey and roadmap
Large Language Models (LLMs), built upon Transformer-based architectures with massive
pretraining on diverse data, have not only revolutionized natural language processing but …
pretraining on diverse data, have not only revolutionized natural language processing but …
Masactrl: Tuning-free mutual self-attention control for consistent image synthesis and editing
Despite the success in large-scale text-to-image generation and text-conditioned image
editing, existing methods still struggle to produce consistent generation and editing results …
editing, existing methods still struggle to produce consistent generation and editing results …
T2i-compbench: A comprehensive benchmark for open-world compositional text-to-image generation
Despite the stunning ability to generate high-quality images by recent text-to-image models,
current approaches often struggle to effectively compose objects with different attributes and …
current approaches often struggle to effectively compose objects with different attributes and …
Layoutgpt: Compositional visual planning and generation with large language models
Attaining a high degree of user controllability in visual generation often requires intricate,
fine-grained inputs like layouts. However, such inputs impose a substantial burden on users …
fine-grained inputs like layouts. However, such inputs impose a substantial burden on users …
Svdiff: Compact parameter space for diffusion fine-tuning
Recently, diffusion models have achieved remarkable success in text-to-image generation,
enabling the creation of high-quality images from text prompts and various conditions …
enabling the creation of high-quality images from text prompts and various conditions …
Fastcomposer: Tuning-free multi-subject image generation with localized attention
Diffusion models excel at text-to-image generation, especially in subject-driven generation
for personalized images. However, existing methods are inefficient due to the subject …
for personalized images. However, existing methods are inefficient due to the subject …
Multimodal foundation models: From specialists to general-purpose assistants
Neural compression is the application of neural networks and other machine learning
methods to data compression. Recent advances in statistical machine learning have opened …
methods to data compression. Recent advances in statistical machine learning have opened …
Boxdiff: Text-to-image synthesis with training-free box-constrained diffusion
Recent text-to-image diffusion models have demonstrated an astonishing capacity to
generate high-quality images. However, researchers mainly studied the way of synthesizing …
generate high-quality images. However, researchers mainly studied the way of synthesizing …
Tokenflow: Consistent diffusion features for consistent video editing
The generative AI revolution has recently expanded to videos. Nevertheless, current state-of-
the-art video models are still lagging behind image models in terms of visual quality and …
the-art video models are still lagging behind image models in terms of visual quality and …
A survey on generative diffusion models
Deep generative models have unlocked another profound realm of human creativity. By
capturing and generalizing patterns within data, we have entered the epoch of all …
capturing and generalizing patterns within data, we have entered the epoch of all …