A review of adversarial attack and defense for classification methods

Y Li, M Cheng, CJ Hsieh, TCM Lee - The American Statistician, 2022 - Taylor & Francis
Despite the efficiency and scalability of machine learning systems, recent studies have
demonstrated that many classification methods, especially Deep Neural Networks (DNNs) …

Better diffusion models further improve adversarial training

Z Wang, T Pang, C Du, M Lin… - … on Machine Learning, 2023 - proceedings.mlr.press
It has been recognized that the data generated by the denoising diffusion probabilistic
model (DDPM) improves adversarial training. After two years of rapid development in …

Cross-entropy loss functions: Theoretical analysis and applications

A Mao, M Mohri, Y Zhong - International conference on …, 2023 - proceedings.mlr.press
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …

Diffusion models for adversarial purification

W Nie, B Guo, Y Huang, C **ao, A Vahdat… - arxiv preprint arxiv …, 2022 - arxiv.org
Adversarial purification refers to a class of defense methods that remove adversarial
perturbations using a generative model. These methods do not make assumptions on the …

Towards trustworthy and aligned machine learning: A data-centric survey with causality perspectives

H Liu, M Chaudhary, H Wang - arxiv preprint arxiv:2307.16851, 2023 - arxiv.org
The trustworthiness of machine learning has emerged as a critical topic in the field,
encompassing various applications and research areas such as robustness, security …

Improving robustness using generated data

S Gowal, SA Rebuffi, O Wiles… - Advances in …, 2021 - proceedings.neurips.cc
Recent work argues that robust training requires substantially larger datasets than those
required for standard classification. On CIFAR-10 and CIFAR-100, this translates into a …

Data augmentation can improve robustness

SA Rebuffi, S Gowal, DA Calian… - Advances in …, 2021 - proceedings.neurips.cc
Adversarial training suffers from robust overfitting, a phenomenon where the robust test
accuracy starts to decrease during training. In this paper, we focus on reducing robust …

Accuracy on the line: on the strong correlation between out-of-distribution and in-distribution generalization

JP Miller, R Taori, A Raghunathan… - International …, 2021 - proceedings.mlr.press
For machine learning systems to be reliable, we must understand their performance in
unseen, out-of-distribution environments. In this paper, we empirically show that out-of …

LAS-AT: adversarial training with learnable attack strategy

X Jia, Y Zhang, B Wu, K Ma… - Proceedings of the …, 2022 - openaccess.thecvf.com
Adversarial training (AT) is always formulated as a minimax problem, of which the
performance depends on the inner optimization that involves the generation of adversarial …

A universal law of robustness via isoperimetry

S Bubeck, M Sellke - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Classically, data interpolation with a parametrized model class is possible as long as the
number of parameters is larger than the number of equations to be satisfied. A puzzling …