Fine-tuning large language models with user-level differential privacy

Z Charles, A Ganesh, R McKenna… - arxiv preprint arxiv …, 2024 - arxiv.org
We investigate practical and scalable algorithms for training large language models (LLMs)
with user-level differential privacy (DP) in order to provably safeguard all the examples …

Subject membership inference attacks in federated learning

A Suri, P Kanani, VJ Marathe, DW Peterson - arxiv preprint arxiv …, 2022 - arxiv.org
Privacy attacks on Machine Learning (ML) models often focus on inferring the existence of
particular data points in the training data. However, what the adversary really wants to know …

User inference attacks on large language models

N Kandpal, K Pillutla, A Oprea, P Kairouz… - arxiv preprint arxiv …, 2023 - arxiv.org
Fine-tuning is a common and effective method for tailoring large language models (LLMs) to
specialized tasks and applications. In this paper, we study the privacy implications of fine …

A General Framework for Data-Use Auditing of ML Models

Z Huang, NZ Gong, MK Reiter - Proceedings of the 2024 on ACM …, 2024 - dl.acm.org
Auditing the use of data in training machine-learning (ML) models is an increasingly
pressing challenge, as myriad ML practitioners routinely leverage the effort of content …

Is my data in your ai model? membership inference test with application to face images

D DeAlcala, A Morales, J Fierrez, G Mancera… - arxiv preprint arxiv …, 2024 - arxiv.org
This article introduces the Membership Inference Test (MINT), a novel approach that aims to
empirically assess if given data was used during the training of AI/ML models. Specifically …

[PDF][PDF] WIP: Auditing Artist Style Pirate in Text-to-image Generation Models

L Du, Z Zhu, M Chen, S Ji, P Cheng… - Proceedings of the …, 2024 - ndss-symposium.org
The text-to-image models based on diffusion processes, capable of transforming text
descriptions into detailed images, have widespread applications in art, design, and beyond …

ORL-AUDITOR: Dataset Auditing in Offline Deep Reinforcement Learning

L Du, M Chen, M Sun, S Ji, P Cheng, J Chen… - arxiv preprint arxiv …, 2023 - arxiv.org
Data is a critical asset in AI, as high-quality datasets can significantly improve the
performance of machine learning models. In safety-critical domains such as autonomous …

SLMIA-SR: Speaker-level membership inference attacks against speaker recognition systems

G Chen, Y Zhang, F Song - arxiv preprint arxiv:2309.07983, 2023 - arxiv.org
Membership inference attacks allow adversaries to determine whether a particular example
was contained in the model's training dataset. While previous works have confirmed the …

Dpmlbench: Holistic evaluation of differentially private machine learning

C Wei, M Zhao, Z Zhang, M Chen, W Meng… - Proceedings of the …, 2023 - dl.acm.org
Differential privacy (DP), as a rigorous mathematical definition quantifying privacy leakage,
has become a well-accepted standard for privacy protection. Combined with powerful …

Range Membership Inference Attacks

J Tao, R Shokri - arxiv preprint arxiv:2408.05131, 2024 - arxiv.org
Machine learning models can leak private information about their training data, but the
standard methods to measure this risk, based on membership inference attacks (MIAs), have …