Interpreting adversarial examples in deep learning: A review

S Han, C Lin, C Shen, Q Wang, X Guan - ACM Computing Surveys, 2023 - dl.acm.org
Deep learning technology is increasingly being applied in safety-critical scenarios but has
recently been found to be susceptible to imperceptible adversarial perturbations. This raises …

The effectiveness of feature attribution methods and its correlation with automatic evaluation scores

G Nguyen, D Kim, A Nguyen - Advances in Neural …, 2021 - proceedings.neurips.cc
Explaining the decisions of an Artificial Intelligence (AI) model is increasingly critical in many
real-world, high-stake applications. Hundreds of papers have either proposed new feature …

How to Defend and Secure Deep Learning Models Against Adversarial Attacks in Computer Vision: A Systematic Review

L Dhamija, U Bansal - New Generation Computing, 2024 - Springer
Deep learning plays a significant role in develo** a robust and constructive framework for
tackling complex learning tasks. Consequently, it is widely utilized in many security-critical …

Adversarial attacks and defenses using feature-space stochasticity

J Ukita, K Ohki - Neural Networks, 2023 - Elsevier
Recent studies in deep neural networks have shown that injecting random noise in the input
layer of the networks contributes towards ℓ p-norm-bounded adversarial perturbations …

Explaining image classifiers by removing input features using generative models

C Agarwal, A Nguyen - Proceedings of the Asian …, 2020 - openaccess.thecvf.com
Perturbation-based explanation methods often measure the contribution of an input feature
to an image classifier's outputs by heuristically removing it via eg blurring, adding noise, or …

Integer-arithmetic-only certified robustness for quantized neural networks

H Lin, J Lou, L **ong… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Adversarial data examples have drawn significant attention from the machine learning and
security communities. A line of work on tackling adversarial examples is certified robustness …

Evaluating the robustness of bayesian neural networks against different types of attacks

Y Pang, S Cheng, J Hu, Y Liu - arxiv preprint arxiv:2106.09223, 2021 - arxiv.org
To evaluate the robustness gain of Bayesian neural networks on image classification tasks,
we perform input perturbations, and adversarial attacks to the state-of-the-art Bayesian …

Ensemble Adversarial Defense via Integration of Multiple Dispersed Low Curvature Models

K Zhao, X Chen, W Huang, L Ding, X Kong… - arxiv preprint arxiv …, 2024 - arxiv.org
The integration of an ensemble of deep learning models has been extensively explored to
enhance defense against adversarial attacks. The diversity among sub-models increases …

On the Trade-offs between Adversarial Robustness and Actionable Explanations

S Krishna, C Agarwal, H Lakkaraju - … of the AAAI/ACM Conference on …, 2024 - ojs.aaai.org
As machine learning models are increasingly being employed in various high-stakes
settings, it becomes important to ensure that predictions of these models are not only …

Robust satellite image classification with Bayesian deep learning

Y Pang, S Cheng, J Hu, Y Liu - 2022 Integrated Communication …, 2022 - ieeexplore.ieee.org
Image-based object detection and classification are essential for satellite-based monitoring,
which spans multiple safety-critical engineering applications. Meanwhile, state-of-the-art …