Wild patterns reloaded: A survey of machine learning security against training data poisoning
The success of machine learning is fueled by the increasing availability of computing power
and large training datasets. The training data is used to learn new models or update existing …
and large training datasets. The training data is used to learn new models or update existing …
Threats, attacks, and defenses in machine unlearning: A survey
Machine Unlearning (MU) has recently gained considerable attention due to its potential to
achieve Safe AI by removing the influence of specific data from trained Machine Learning …
achieve Safe AI by removing the influence of specific data from trained Machine Learning …
Backdoor learning: A survey
Backdoor attack intends to embed hidden backdoors into deep neural networks (DNNs), so
that the attacked models perform well on benign samples, whereas their predictions will be …
that the attacked models perform well on benign samples, whereas their predictions will be …
Narcissus: A practical clean-label backdoor attack with limited information
Backdoor attacks introduce manipulated data into a machine learning model's training set,
causing the model to misclassify inputs with a trigger during testing to achieve a desired …
causing the model to misclassify inputs with a trigger during testing to achieve a desired …
Backdoor defense via decoupling the training process
Recent studies have revealed that deep neural networks (DNNs) are vulnerable to backdoor
attacks, where attackers embed hidden backdoors in the DNN model by poisoning a few …
attacks, where attackers embed hidden backdoors in the DNN model by poisoning a few …
Backdoor defense via adaptively splitting poisoned dataset
Backdoor defenses have been studied to alleviate the threat of deep neural networks
(DNNs) being backdoor attacked and thus maliciously altered. Since DNNs usually adopt …
(DNNs) being backdoor attacked and thus maliciously altered. Since DNNs usually adopt …
Neural polarizer: A lightweight and effective backdoor defense via purifying poisoned features
Recent studies have demonstrated the susceptibility of deep neural networks to backdoor
attacks. Given a backdoored model, its prediction of a poisoned sample with trigger will be …
attacks. Given a backdoored model, its prediction of a poisoned sample with trigger will be …
Shared adversarial unlearning: Backdoor mitigation by unlearning shared adversarial examples
Backdoor attacks are serious security threats to machine learning models where an
adversary can inject poisoned samples into the training set, causing a backdoored model …
adversary can inject poisoned samples into the training set, causing a backdoored model …
Mm-bd: Post-training detection of backdoor attacks with arbitrary backdoor pattern types using a maximum margin statistic
Backdoor attacks are an important type of adversarial threat against deep neural network
classifiers, wherein test samples from one or more source classes will be (mis) classified to …
classifiers, wherein test samples from one or more source classes will be (mis) classified to …
Rethinking the reverse-engineering of trojan triggers
Abstract Deep Neural Networks are vulnerable to Trojan (or backdoor) attacks. Reverse-
engineering methods can reconstruct the trigger and thus identify affected models. Existing …
engineering methods can reconstruct the trigger and thus identify affected models. Existing …