Membership inference attacks on machine learning: A survey
Machine learning (ML) models have been widely applied to various applications, including
image classification, text generation, audio recognition, and graph data analysis. However …
image classification, text generation, audio recognition, and graph data analysis. However …
Ammus: A survey of transformer-based pretrained models in natural language processing
KS Kalyan, A Rajasekharan, S Sangeetha - arxiv preprint arxiv …, 2021 - arxiv.org
Transformer-based pretrained language models (T-PTLMs) have achieved great success in
almost every NLP task. The evolution of these models started with GPT and BERT. These …
almost every NLP task. The evolution of these models started with GPT and BERT. These …
Extracting training data from large language models
It has become common to publish large (billion parameter) language models that have been
trained on private datasets. This paper demonstrates that in such settings, an adversary can …
trained on private datasets. This paper demonstrates that in such settings, an adversary can …
Red teaming language models with language models
Language Models (LMs) often cannot be deployed because of their potential to harm users
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …
Propile: Probing privacy leakage in large language models
The rapid advancement and widespread use of large language models (LLMs) have raised
significant concerns regarding the potential leakage of personally identifiable information …
significant concerns regarding the potential leakage of personally identifiable information …
A survey of privacy attacks in machine learning
As machine learning becomes more widely used, the need to study its implications in
security and privacy becomes more urgent. Although the body of work in privacy has been …
security and privacy becomes more urgent. Although the body of work in privacy has been …
Membership inference attacks against language models via neighbourhood comparison
Membership Inference attacks (MIAs) aim to predict whether a data sample was present in
the training data of a machine learning model or not, and are widely used for assessing the …
the training data of a machine learning model or not, and are widely used for assessing the …
A study of face obfuscation in imagenet
Face obfuscation (blurring, mosaicing, etc.) has been shown to be effective for privacy
protection; nevertheless, object recognition research typically assumes access to complete …
protection; nevertheless, object recognition research typically assumes access to complete …
Beyond the safeguards: exploring the security risks of ChatGPT
E Derner, K Batistič - arxiv preprint arxiv:2305.08005, 2023 - arxiv.org
The increasing popularity of large language models (LLMs) such as ChatGPT has led to
growing concerns about their safety, security risks, and ethical implications. This paper aims …
growing concerns about their safety, security risks, and ethical implications. This paper aims …
Quantifying privacy risks of masked language models using membership inference attacks
The wide adoption and application of Masked language models~(MLMs) on sensitive data
(from legal to medical) necessitates a thorough quantitative investigation into their privacy …
(from legal to medical) necessitates a thorough quantitative investigation into their privacy …