Wild patterns reloaded: A survey of machine learning security against training data poisoning
The success of machine learning is fueled by the increasing availability of computing power
and large training datasets. The training data is used to learn new models or update existing …
and large training datasets. The training data is used to learn new models or update existing …
Backdoor learning: A survey
Backdoor attack intends to embed hidden backdoors into deep neural networks (DNNs), so
that the attacked models perform well on benign samples, whereas their predictions will be …
that the attacked models perform well on benign samples, whereas their predictions will be …
Anti-backdoor learning: Training clean models on poisoned data
Backdoor attack has emerged as a major security threat to deep neural networks (DNNs).
While existing defense methods have demonstrated promising results on detecting or …
While existing defense methods have demonstrated promising results on detecting or …
An overview of backdoor attacks against deep neural networks and possible defences
Together with impressive advances touching every aspect of our society, AI technology
based on Deep Neural Networks (DNN) is bringing increasing security concerns. While …
based on Deep Neural Networks (DNN) is bringing increasing security concerns. While …
Bppattack: Stealthy and efficient trojan attacks against deep neural networks via image quantization and contrastive adversarial learning
Deep neural networks are vulnerable to Trojan attacks. Existing attacks use visible patterns
(eg, a patch or image transformations) as triggers, which are vulnerable to human …
(eg, a patch or image transformations) as triggers, which are vulnerable to human …
Revisiting the assumption of latent separability for backdoor defenses
Recent studies revealed that deep learning is susceptible to backdoor poisoning attacks. An
adversary can embed a hidden backdoor into a model to manipulate its predictions by only …
adversary can embed a hidden backdoor into a model to manipulate its predictions by only …
Color backdoor: A robust poisoning attack in color space
Backdoor attacks against neural networks have been intensively investigated, where the
adversary compromises the integrity of the victim model, causing it to make wrong …
adversary compromises the integrity of the victim model, causing it to make wrong …
Defeat: Deep hidden feature backdoor attacks by imperceptible perturbation and latent representation constraints
Backdoor attack is a type of serious security threat to deep learning models. An adversary
can provide users with a model trained on poisoned data to manipulate prediction behavior …
can provide users with a model trained on poisoned data to manipulate prediction behavior …
Detecting backdoors in pre-trained encoders
Self-supervised learning in computer vision trains on unlabeled data, such as images or
(image, text) pairs, to obtain an image encoder that learns high-quality embeddings for input …
(image, text) pairs, to obtain an image encoder that learns high-quality embeddings for input …
Hidden backdoors in human-centric language models
Natural language processing (NLP) systems have been proven to be vulnerable to backdoor
attacks, whereby hidden features (backdoors) are trained into a language model and may …
attacks, whereby hidden features (backdoors) are trained into a language model and may …