[HTML][HTML] A survey on large language model (llm) security and privacy: The good, the bad, and the ugly
Abstract Large Language Models (LLMs), such as ChatGPT and Bard, have revolutionized
natural language understanding and generation. They possess deep language …
natural language understanding and generation. They possess deep language …
Membership inference attacks on machine learning: A survey
Machine learning (ML) models have been widely applied to various applications, including
image classification, text generation, audio recognition, and graph data analysis. However …
image classification, text generation, audio recognition, and graph data analysis. However …
Scalable extraction of training data from (production) language models
This paper studies extractable memorization: training data that an adversary can efficiently
extract by querying a machine learning model without prior knowledge of the training …
extract by querying a machine learning model without prior knowledge of the training …
Membership inference attacks from first principles
A membership inference attack allows an adversary to query a trained machine learning
model to predict whether or not a particular example was contained in the model's training …
model to predict whether or not a particular example was contained in the model's training …
Trustworthy LLMs: A survey and guideline for evaluating large language models' alignment
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …
Are diffusion models vulnerable to membership inference attacks?
Diffusion-based generative models have shown great potential for image synthesis, but
there is a lack of research on the security and privacy risks they may pose. In this paper, we …
there is a lack of research on the security and privacy risks they may pose. In this paper, we …
Enhanced membership inference attacks against machine learning models
How much does a machine learning algorithm leak about its training data, and why?
Membership inference attacks are used as an auditing tool to quantify this leakage. In this …
Membership inference attacks are used as an auditing tool to quantify this leakage. In this …
Threats, attacks and defenses to federated learning: issues, taxonomy and perspectives
Abstract Empirical attacks on Federated Learning (FL) systems indicate that FL is fraught
with numerous attack surfaces throughout the FL execution. These attacks can not only …
with numerous attack surfaces throughout the FL execution. These attacks can not only …
Membership inference attacks against language models via neighbourhood comparison
Membership Inference attacks (MIAs) aim to predict whether a data sample was present in
the training data of a machine learning model or not, and are widely used for assessing the …
the training data of a machine learning model or not, and are widely used for assessing the …
A survey of privacy attacks in machine learning
As machine learning becomes more widely used, the need to study its implications in
security and privacy becomes more urgent. Although the body of work in privacy has been …
security and privacy becomes more urgent. Although the body of work in privacy has been …