Trustworthy AI: From principles to practices

B Li, P Qi, B Liu, S Di, J Liu, J Pei, J Yi… - ACM Computing Surveys, 2023 - dl.acm.org
The rapid development of Artificial Intelligence (AI) technology has enabled the deployment
of various systems based on it. However, many current AI systems are found vulnerable to …

Adversarial attacks and defenses in deep learning: From a perspective of cybersecurity

S Zhou, C Liu, D Ye, T Zhu, W Zhou, PS Yu - ACM Computing Surveys, 2022 - dl.acm.org
The outstanding performance of deep neural networks has promoted deep learning
applications in a broad set of domains. However, the potential risks caused by adversarial …

Adversarial examples are not bugs, they are features

A Ilyas, S Santurkar, D Tsipras… - Advances in neural …, 2019 - proceedings.neurips.cc
Adversarial examples have attracted significant attention in machine learning, but the
reasons for their existence and pervasiveness remain unclear. We demonstrate that …

Theoretically principled trade-off between robustness and accuracy

H Zhang, Y Yu, J Jiao, E **ng… - International …, 2019 - proceedings.mlr.press
We identify a trade-off between robustness and accuracy that serves as a guiding principle
in the design of defenses against adversarial examples. Although this problem has been …

Robustness may be at odds with accuracy

D Tsipras, S Santurkar, L Engstrom, A Turner… - arxiv preprint arxiv …, 2018 - arxiv.org
We show that there may exist an inherent tension between the goal of adversarial
robustness and that of standard generalization. Specifically, training robust models may not …

Adversarial examples: Attacks and defenses for deep learning

X Yuan, P He, Q Zhu, X Li - IEEE transactions on neural …, 2019 - ieeexplore.ieee.org
With rapid progress and significant successes in a wide spectrum of applications, deep
learning is being applied in many safety-critical environments. However, deep neural …

Simple black-box adversarial attacks

C Guo, J Gardner, Y You, AG Wilson… - … on machine learning, 2019 - proceedings.mlr.press
We propose an intriguingly simple method for the construction of adversarial images in the
black-box setting. In constrast to the white-box scenario, constructing black-box adversarial …

Adversarially robust generalization requires more data

L Schmidt, S Santurkar, D Tsipras… - Advances in neural …, 2018 - proceedings.neurips.cc
Abstract Machine learning models are often susceptible to adversarial perturbations of their
inputs. Even small perturbations can cause state-of-the-art classifiers with high" standard" …

Adversarial training and robustness for multiple perturbations

F Tramer, D Boneh - Advances in neural information …, 2019 - proceedings.neurips.cc
Defenses against adversarial examples, such as adversarial training, are typically tailored to
a single perturbation type (eg, small $\ell_\infty $-noise). For other perturbations, these …

A closer look at accuracy vs. robustness

YY Yang, C Rashtchian, H Zhang… - Advances in neural …, 2020 - proceedings.neurips.cc
Current methods for training robust networks lead to a drop in test accuracy, which has led
prior works to posit that a robustness-accuracy tradeoff may be inevitable in deep learning …