Counterfactual explanations and how to find them: literature review and benchmarking

R Guidotti - Data Mining and Knowledge Discovery, 2024 - Springer
Interpretable machine learning aims at unveiling the reasons behind predictions returned by
uninterpretable classifiers. One of the most valuable types of explanation consists of …

A review on generative adversarial networks: Algorithms, theory, and applications

J Gui, Z Sun, Y Wen, D Tao, J Ye - IEEE transactions on …, 2021 - ieeexplore.ieee.org
Generative adversarial networks (GANs) have recently become a hot research topic;
however, they have been studied since 2014, and a large number of algorithms have been …

Adding conditional control to text-to-image diffusion models

L Zhang, A Rao, M Agrawala - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
We present ControlNet, a neural network architecture to add spatial conditioning controls to
large, pretrained text-to-image diffusion models. ControlNet locks the production-ready large …

Palette: Image-to-image diffusion models

C Saharia, W Chan, H Chang, C Lee, J Ho… - ACM SIGGRAPH 2022 …, 2022 - dl.acm.org
This paper develops a unified framework for image-to-image translation based on
conditional diffusion models and evaluates this framework on four challenging image-to …

Zero-shot image-to-image translation

G Parmar, K Kumar Singh, R Zhang, Y Li, J Lu… - ACM SIGGRAPH 2023 …, 2023 - dl.acm.org
Large-scale text-to-image generative models have shown their remarkable ability to
synthesize diverse, high-quality images. However, directly applying these models for real …

Diffusionclip: Text-guided diffusion models for robust image manipulation

G Kim, T Kwon, JC Ye - … of the IEEE/CVF conference on …, 2022 - openaccess.thecvf.com
Recently, GAN inversion methods combined with Contrastive Language-Image Pretraining
(CLIP) enables zero-shot image manipulation guided by text prompts. However, their …

Contrastive learning for unpaired image-to-image translation

T Park, AA Efros, R Zhang, JY Zhu - … , Glasgow, UK, August 23–28, 2020 …, 2020 - Springer
In image-to-image translation, each patch in the output should reflect the content of the
corresponding patch in the input, independent of domain. We propose a straightforward …

Giraffe: Representing scenes as compositional generative neural feature fields

M Niemeyer, A Geiger - … of the IEEE/CVF Conference on …, 2021 - openaccess.thecvf.com
Deep generative models allow for photorealistic image synthesis at high resolutions. But for
many applications, this is not enough: content creation also needs to be controllable. While …

Gligen: Open-set grounded text-to-image generation

Y Li, H Liu, Q Wu, F Mu, J Yang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale text-to-image diffusion models have made amazing advances. However, the
status quo is to use text input alone, which can impede controllability. In this work, we …

Image fusion in the loop of high-level vision tasks: A semantic-aware real-time infrared and visible image fusion network

L Tang, J Yuan, J Ma - Information Fusion, 2022 - Elsevier
Infrared and visible image fusion aims to synthesize a single fused image that not only
contains salient targets and abundant texture details but also facilitates high-level vision …