Diffusion models: A comprehensive survey of methods and applications

L Yang, Z Zhang, Y Song, S Hong, R Xu, Y Zhao… - ACM Computing …, 2023 - dl.acm.org
Diffusion models have emerged as a powerful new family of deep generative models with
record-breaking performance in many applications, including image synthesis, video …

Smoothllm: Defending large language models against jailbreaking attacks

A Robey, E Wong, H Hassani, GJ Pappas - arxiv preprint arxiv …, 2023 - arxiv.org
Despite efforts to align large language models (LLMs) with human values, widely-used
LLMs such as GPT, Llama, Claude, and PaLM are susceptible to jailbreaking attacks …

Masked autoencoders in computer vision: A comprehensive survey

Z Zhou, X Liu - IEEE Access, 2023 - ieeexplore.ieee.org
Masked autoencoders (MAE) is a deep learning method based on Transformer. Originally
used for images, it has now been extended to video, audio, and some other temporal …

Single image backdoor inversion via robust smoothed classifiers

M Sun, Z Kolter - Proceedings of the IEEE/CVF Conference …, 2023 - openaccess.thecvf.com
Backdoor inversion, the process of finding a backdoor trigger inserted into a machine
learning model, has become the pillar of many backdoor detection and defense methods …

Improving the accuracy-robustness trade-off of classifiers via adaptive smoothing

Y Bai, BG Anderson, A Kim, S Sojoudi - SIAM Journal on Mathematics of Data …, 2024 - SIAM
While prior research has proposed a plethora of methods that build neural classifiers robust
against adversarial robustness, practitioners are still reluctant to adopt them due to their …

Towards Universal Detection of Adversarial Examples via Pseudorandom Classifiers

B Zhu, C Dong, Y Zhang, Y Mao… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Adversarial examples that can fool neural network classifiers have attracted much attention.
Existing approaches to detect adversarial examples leverage a supervised scheme in …

Destruction-Restoration Suppresses Data Protection Perturbations against Diffusion Models

T Qin, X Gao, J Zhao, K Ye - 2023 IEEE 35th International …, 2023 - ieeexplore.ieee.org
Diffusion models have become popular in computer vision applications due to their ability to
generate high-quality images quickly, with some models achieving a high degree of realism …

Boosting certified robustness via an expectation-based similarity regularization

J Li, K Fang, X Huang, J Yang - Image and Vision Computing, 2024 - Elsevier
A certifiably robust classifier implies the one that is theoretically guaranteed to provide
robust predictions against any adversarial attacks under certain conditions. Recent defense …

Purify++: Improving Diffusion-Purification with Advanced Diffusion Models and Control of Randomness

B Zhang, W Luo, Z Zhang - arxiv preprint arxiv:2310.18762, 2023 - arxiv.org
Adversarial attacks can mislead neural network classifiers. The defense against adversarial
attacks is important for AI safety. Adversarial purification is a family of approaches that …

Case Study: Neural Network Malware Detection Verification for Feature and Image Datasets

PK Robinette, D Manzanas Lopez… - Proceedings of the …, 2024 - dl.acm.org
Malware, or software designed with harmful intent, is an ever-evolving threat that can have
drastic effects on both individuals and institutions. Neural network malware classification …