Backdoor attacks and countermeasures on deep learning: A comprehensive review

Y Gao, BG Doan, Z Zhang, S Ma, J Zhang, A Fu… - arxiv preprint arxiv …, 2020 - arxiv.org
This work provides the community with a timely comprehensive review of backdoor attacks
and countermeasures on deep learning. According to the attacker's capability and affected …

Backdoor attacks and defenses targeting multi-domain ai models: A comprehensive review

S Zhang, Y Pan, Q Liu, Z Yan, KKR Choo… - ACM Computing …, 2024 - dl.acm.org
Since the emergence of security concerns in artificial intelligence (AI), there has been
significant attention devoted to the examination of backdoor attacks. Attackers can utilize …

Toward transparent ai: A survey on interpreting the inner structures of deep neural networks

T Räuker, A Ho, S Casper… - 2023 ieee conference …, 2023 - ieeexplore.ieee.org
The last decade of machine learning has seen drastic increases in scale and capabilities.
Deep neural networks (DNNs) are increasingly being deployed in the real world. However …

Untargeted backdoor watermark: Towards harmless and stealthy dataset copyright protection

Y Li, Y Bai, Y Jiang, Y Yang… - Advances in Neural …, 2022 - proceedings.neurips.cc
Deep neural networks (DNNs) have demonstrated their superiority in practice. Arguably, the
rapid development of DNNs is largely benefited from high-quality (open-sourced) datasets …

Februus: Input purification defense against trojan attacks on deep neural network systems

BG Doan, E Abbasnejad, DC Ranasinghe - Proceedings of the 36th …, 2020 - dl.acm.org
We propose Februus; a new idea to neutralize highly potent and insidious Trojan attacks on
Deep Neural Network (DNN) systems at run-time. In Trojan attacks, an adversary activates a …

A unified evaluation of textual backdoor learning: Frameworks and benchmarks

G Cui, L Yuan, B He, Y Chen… - Advances in Neural …, 2022 - proceedings.neurips.cc
Textual backdoor attacks are a kind of practical threat to NLP systems. By injecting a
backdoor in the training phase, the adversary could control model predictions via predefined …

Not all samples are born equal: Towards effective clean-label backdoor attacks

Y Gao, Y Li, L Zhu, D Wu, Y Jiang, ST **a - Pattern Recognition, 2023 - Elsevier
Recent studies demonstrated that deep neural networks (DNNs) are vulnerable to backdoor
attacks. The attacked model behaves normally on benign samples, while its predictions are …

Scale-up: An efficient black-box input-level backdoor detection via analyzing scaled prediction consistency

J Guo, Y Li, X Chen, H Guo, L Sun, C Liu - arxiv preprint arxiv:2302.03251, 2023 - arxiv.org
Deep neural networks (DNNs) are vulnerable to backdoor attacks, where adversaries
embed a hidden backdoor trigger during the training process for malicious prediction …

Rap: Robustness-aware perturbations for defending against backdoor attacks on nlp models

W Yang, Y Lin, P Li, J Zhou, X Sun - arxiv preprint arxiv:2110.07831, 2021 - arxiv.org
Backdoor attacks, which maliciously control a well-trained model's outputs of the instances
with specific triggers, are recently shown to be serious threats to the safety of reusing deep …

Can we use split learning on 1D CNN models for privacy preserving training?

S Abuadbba, K Kim, M Kim, C Thapa… - Proceedings of the 15th …, 2020 - dl.acm.org
A new collaborative learning, called split learning, was recently introduced, aiming to protect
user data privacy without revealing raw input data to a server. It collaboratively runs a deep …