To generate or not? safety-driven unlearned diffusion models are still easy to generate unsafe images... for now

Y Zhang, J Jia, X Chen, A Chen, Y Zhang, J Liu… - … on Computer Vision, 2024 - Springer
The recent advances in diffusion models (DMs) have revolutionized the generation of
realistic and complex images. However, these models also introduce potential safety …

Defensive unlearning with adversarial training for robust concept erasure in diffusion models

Y Zhang, X Chen, J Jia, Y Zhang… - Advances in …, 2025 - proceedings.neurips.cc
Diffusion models (DMs) have achieved remarkable success in text-to-image generation, but
they also pose safety risks, such as the potential generation of harmful content and copyright …

Visual prompting for adversarial robustness

A Chen, P Lorenz, Y Yao, PY Chen… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
In this work, we leverage visual prompting (VP) to improve adversarial robustness of a fixed,
pre-trained model at test time. Compared to conventional adversarial defenses, VP allows …

Reverse engineering of deceptions on machine-and human-centric attacks

Y Yao, X Guo, V Asnani, Y Gong, J Liu… - … and Trends® in …, 2024 - nowpublishers.com
This work presents a comprehensive exploration of Reverse Engineering of Deceptions
(RED) in the field of adversarial machine learning. It delves into the intricacies of machine …

Improving adversarial robustness of medical imaging systems via adding global attention noise

Y Dai, Y Qian, F Lu, B Wang, Z Gu, W Wang… - Computers in Biology …, 2023 - Elsevier
Recent studies have found that medical images are vulnerable to adversarial attacks.
However, it is difficult to protect medical imaging systems from adversarial examples in that …

Holistic adversarial robustness of deep learning models

PY Chen, S Liu - Proceedings of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
Adversarial robustness studies the worst-case performance of a machine learning model to
ensure safety and reliability. With the proliferation of deep-learning-based technology, the …

Less is more: Data pruning for faster adversarial training

Y Li, P Zhao, X Lin, B Kailkhura, R Goldhahn - arxiv preprint arxiv …, 2023 - arxiv.org
Deep neural networks (DNNs) are sensitive to adversarial examples, resulting in fragile and
unreliable performance in the real world. Although adversarial training (AT) is currently one …

Neural architecture search for adversarial robustness via learnable pruning

Y Li, P Zhao, R Ding, T Zhou, Y Fei, X Xu… - Frontiers in High …, 2024 - frontiersin.org
The convincing performances of deep neural networks (DNNs) can be degraded
tremendously under malicious samples, known as adversarial examples. Besides, with the …

Uncovering Distortion Differences: A Study of Adversarial Attacks and Machine Discriminability

X Wang, Y Li, CJ Hsieh, TCM Lee - IEEE Access, 2024 - ieeexplore.ieee.org
Deep neural networks have performed remarkably in many areas, including image-related
classification tasks. However, various studies have shown that they are vulnerable to …

Tracing hyperparameter dependencies for model parsing via learnable graph pooling network

X Guo, V Asnani, S Liu, X Liu - arxiv preprint arxiv:2312.02224, 2023 - arxiv.org
Model Parsing defines the research task of predicting hyperparameters of the generative
model (GM), given a generated image as input. Since a diverse set of hyperparameters is …