A comprehensive survey on poisoning attacks and countermeasures in machine learning

Z Tian, L Cui, J Liang, S Yu - ACM Computing Surveys, 2022 - dl.acm.org
The prosperity of machine learning has been accompanied by increasing attacks on the
training process. Among them, poisoning attacks have become an emerging threat during …

Generative adversarial networks: A survey toward private and secure applications

Z Cai, Z **ong, H Xu, P Wang, W Li, Y Pan - ACM Computing Surveys …, 2021 - dl.acm.org
Generative Adversarial Networks (GANs) have promoted a variety of applications in
computer vision and natural language processing, among others, due to its generative …

Extracting training data from diffusion models

N Carlini, J Hayes, M Nasr, M Jagielski… - 32nd USENIX Security …, 2023 - usenix.org
Image diffusion models such as DALL-E 2, Imagen, and Stable Diffusion have attracted
significant attention due to their ability to generate high-quality synthetic images. In this work …

Trustworthy llms: a survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, R Guo, H Cheng… - arxiv preprint arxiv …, 2023 - arxiv.org
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …

Propile: Probing privacy leakage in large language models

S Kim, S Yun, H Lee, M Gubri… - Advances in Neural …, 2023 - proceedings.neurips.cc
The rapid advancement and widespread use of large language models (LLMs) have raised
significant concerns regarding the potential leakage of personally identifiable information …

Towards unbounded machine unlearning

M Kurmanji, P Triantafillou, J Hayes… - Advances in neural …, 2023 - proceedings.neurips.cc
Deep machine unlearning is the problem of'removing'from a trained neural network a subset
of its training set. This problem is very timely and has many applications, including the key …

Are diffusion models vulnerable to membership inference attacks?

J Duan, F Kong, S Wang, X Shi… - … Conference on Machine …, 2023 - proceedings.mlr.press
Diffusion-based generative models have shown great potential for image synthesis, but
there is a lack of research on the security and privacy risks they may pose. In this paper, we …

Fedfed: Feature distillation against data heterogeneity in federated learning

Z Yang, Y Zhang, Y Zheng, X Tian… - Advances in …, 2023 - proceedings.neurips.cc
Federated learning (FL) typically faces data heterogeneity, ie, distribution shifting among
clients. Sharing clients' information has shown great potentiality in mitigating data …

See through gradients: Image batch recovery via gradinversion

H Yin, A Mallya, A Vahdat, JM Alvarez… - Proceedings of the …, 2021 - openaccess.thecvf.com
Training deep neural networks requires gradient estimation from data batches to update
parameters. Gradients per parameter are averaged over a set of data and this has been …

Reconstructing training data from trained neural networks

N Haim, G Vardi, G Yehudai… - Advances in Neural …, 2022 - proceedings.neurips.cc
Understanding to what extent neural networks memorize training data is an intriguing
question with practical and theoretical implications. In this paper we show that in some …