Fine-tuning large language models with user-level differential privacy
We investigate practical and scalable algorithms for training large language models (LLMs)
with user-level differential privacy (DP) in order to provably safeguard all the examples …
with user-level differential privacy (DP) in order to provably safeguard all the examples …
Subject membership inference attacks in federated learning
Privacy attacks on Machine Learning (ML) models often focus on inferring the existence of
particular data points in the training data. However, what the adversary really wants to know …
particular data points in the training data. However, what the adversary really wants to know …
User inference attacks on large language models
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to
specialized tasks and applications. In this paper, we study the privacy implications of fine …
specialized tasks and applications. In this paper, we study the privacy implications of fine …
A General Framework for Data-Use Auditing of ML Models
Auditing the use of data in training machine-learning (ML) models is an increasingly
pressing challenge, as myriad ML practitioners routinely leverage the effort of content …
pressing challenge, as myriad ML practitioners routinely leverage the effort of content …
Is my data in your ai model? membership inference test with application to face images
This article introduces the Membership Inference Test (MINT), a novel approach that aims to
empirically assess if given data was used during the training of AI/ML models. Specifically …
empirically assess if given data was used during the training of AI/ML models. Specifically …
[PDF][PDF] WIP: Auditing Artist Style Pirate in Text-to-image Generation Models
The text-to-image models based on diffusion processes, capable of transforming text
descriptions into detailed images, have widespread applications in art, design, and beyond …
descriptions into detailed images, have widespread applications in art, design, and beyond …
ORL-AUDITOR: Dataset Auditing in Offline Deep Reinforcement Learning
Data is a critical asset in AI, as high-quality datasets can significantly improve the
performance of machine learning models. In safety-critical domains such as autonomous …
performance of machine learning models. In safety-critical domains such as autonomous …
SLMIA-SR: Speaker-level membership inference attacks against speaker recognition systems
Membership inference attacks allow adversaries to determine whether a particular example
was contained in the model's training dataset. While previous works have confirmed the …
was contained in the model's training dataset. While previous works have confirmed the …
Dpmlbench: Holistic evaluation of differentially private machine learning
Differential privacy (DP), as a rigorous mathematical definition quantifying privacy leakage,
has become a well-accepted standard for privacy protection. Combined with powerful …
has become a well-accepted standard for privacy protection. Combined with powerful …
Range Membership Inference Attacks
Machine learning models can leak private information about their training data, but the
standard methods to measure this risk, based on membership inference attacks (MIAs), have …
standard methods to measure this risk, based on membership inference attacks (MIAs), have …