Wild patterns reloaded: A survey of machine learning security against training data poisoning

AE Cinà, K Grosse, A Demontis, S Vascon… - ACM Computing …, 2023 - dl.acm.org
The success of machine learning is fueled by the increasing availability of computing power
and large training datasets. The training data is used to learn new models or update existing …

On the exploitability of instruction tuning

M Shu, J Wang, C Zhu, J Gei**… - Advances in Neural …, 2023 - proceedings.neurips.cc
Instruction tuning is an effective technique to align large language models (LLMs) with
human intent. In this work, we investigate how an adversary can exploit instruction tuning by …

Training data influence analysis and estimation: A survey

Z Hammoudeh, D Lowd - Machine Learning, 2024 - Springer
Good models require good training data. For overparameterized deep models, the causal
relationship between training data and model predictions is increasingly opaque and poorly …

Truth serum: Poisoning machine learning models to reveal their secrets

F Tramèr, R Shokri, A San Joaquin, H Le… - Proceedings of the …, 2022 - dl.acm.org
We introduce a new class of attacks on machine learning models. We show that an
adversary who can poison a training dataset can cause models trained on this dataset to …

Unlearnable 3D point clouds: Class-wise transformation is all you need

X Wang, M Li, W Liu, H Zhang, S Hu… - Advances in …, 2025 - proceedings.neurips.cc
Traditional unlearnable strategies have been proposed to prevent unauthorized users from
training on the 2D image data. With more 3D point cloud data containing sensitivity …

Sleeper agent: Scalable hidden trigger backdoors for neural networks trained from scratch

H Souri, L Fowl, R Chellappa… - Advances in …, 2022 - proceedings.neurips.cc
As the curation of data for machine learning becomes increasingly automated, dataset
tampering is a mounting threat. Backdoor attackers tamper with training data to embed a …

The path to defence: A roadmap to characterising data poisoning attacks on victim models

T Chaalan, S Pang, J Kamruzzaman, I Gondal… - ACM Computing …, 2024 - dl.acm.org
Data Poisoning Attacks (DPA) represent a sophisticated technique aimed at distorting the
training data of machine learning models, thereby manipulating their behavior. This process …

Robust unlearnable examples: Protecting data against adversarial learning

S Fu, F He, Y Liu, L Shen, D Tao - arxiv preprint arxiv:2203.14533, 2022 - arxiv.org
The tremendous amount of accessible data in cyberspace face the risk of being
unauthorized used for training deep learning models. To address this concern, methods are …

Cuda: Convolution-based unlearnable datasets

VS Sadasivan, M Soltanolkotabi… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale training of modern deep learning models heavily relies on publicly available
data on the web. This potentially unauthorized usage of online data leads to concerns …

Image shortcut squeezing: Countering perturbative availability poisons with compression

Z Liu, Z Zhao, M Larson - International conference on …, 2023 - proceedings.mlr.press
Perturbative availability poisoning (PAP) adds small changes to images to prevent their use
for model training. Current research adopts the belief that practical and effective approaches …