Membership inference attacks on machine learning: A survey

H Hu, Z Salcic, L Sun, G Dobbie, PS Yu… - ACM Computing Surveys …, 2022 - dl.acm.org
Machine learning (ML) models have been widely applied to various applications, including
image classification, text generation, audio recognition, and graph data analysis. However …

Ammus: A survey of transformer-based pretrained models in natural language processing

KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …

Extracting training data from large language models

N Carlini, F Tramer, E Wallace, M Jagielski… - 30th USENIX Security …, 2021 - usenix.org
It has become common to publish large (billion parameter) language models that have been
trained on private datasets. This paper demonstrates that in such settings, an adversary can …

Red teaming language models with language models

E Perez, S Huang, F Song, T Cai, R Ring… - arxiv preprint arxiv …, 2022 - arxiv.org
Language Models (LMs) often cannot be deployed because of their potential to harm users
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …

Propile: Probing privacy leakage in large language models

S Kim, S Yun, H Lee, M Gubri… - Advances in Neural …, 2024 - proceedings.neurips.cc
The rapid advancement and widespread use of large language models (LLMs) have raised
significant concerns regarding the potential leakage of personally identifiable information …

A survey of privacy attacks in machine learning

M Rigaki, S Garcia - ACM Computing Surveys, 2023 - dl.acm.org
As machine learning becomes more widely used, the need to study its implications in
security and privacy becomes more urgent. Although the body of work in privacy has been …

Membership inference attacks against language models via neighbourhood comparison

J Mattern, F Mireshghallah, Z **, B Schölkopf… - arxiv preprint arxiv …, 2023 - arxiv.org
Membership Inference attacks (MIAs) aim to predict whether a data sample was present in
the training data of a machine learning model or not, and are widely used for assessing the …

A study of face obfuscation in imagenet

K Yang, JH Yau, L Fei-Fei, J Deng… - International …, 2022 - proceedings.mlr.press
Face obfuscation (blurring, mosaicing, etc.) has been shown to be effective for privacy
protection; nevertheless, object recognition research typically assumes access to complete …

Beyond the safeguards: exploring the security risks of ChatGPT

E Derner, K Batistič - arxiv preprint arxiv:2305.08005, 2023 - arxiv.org
The increasing popularity of large language models (LLMs) such as ChatGPT has led to
growing concerns about their safety, security risks, and ethical implications. This paper aims …

Quantifying privacy risks of masked language models using membership inference attacks

F Mireshghallah, K Goyal, A Uniyal… - arxiv preprint arxiv …, 2022 - arxiv.org
The wide adoption and application of Masked language models~(MLMs) on sensitive data
(from legal to medical) necessitates a thorough quantitative investigation into their privacy …