How to dp-fy ml: A practical guide to machine learning with differential privacy

N Ponomareva, H Hazimeh, A Kurakin, Z Xu… - Journal of Artificial …, 2023 - jair.org
Abstract Machine Learning (ML) models are ubiquitous in real-world applications and are a
constant focus of research. Modern ML models have become more complex, deeper, and …

Privacy auditing with one (1) training run

T Steinke, M Nasr, M Jagielski - Advances in Neural …, 2024 - proceedings.neurips.cc
We propose a scheme for auditing differentially private machine learning systems with a
single training run. This exploits the parallelism of being able to add or remove multiple …

A sco** review of privacy and utility metrics in medical synthetic data

B Kaabachi, J Despraz, T Meurers, K Otte… - NPJ digital …, 2025 - nature.com
The use of synthetic data is a promising solution to facilitate the sharing and reuse of health-
related data beyond its initial collection while addressing privacy concerns. However, there …

Detecting pretraining data from large language models

W Shi, A Ajith, M **a, Y Huang, D Liu, T Blevins… - arxiv preprint arxiv …, 2023 - arxiv.org
Although large language models (LLMs) are widely deployed, the data used to train them is
rarely disclosed. Given the incredible scale of this data, up to trillions of tokens, it is all but …

Label poisoning is all you need

R Jha, J Hayase, S Oh - Advances in Neural Information …, 2023 - proceedings.neurips.cc
In a backdoor attack, an adversary injects corrupted data into a model's training dataset in
order to gain control over its predictions on images with a specific attacker-defined trigger. A …

Evaluations of machine learning privacy defenses are misleading

M Aerni, J Zhang, F Tramèr - Proceedings of the 2024 on ACM SIGSAC …, 2024 - dl.acm.org
Empirical defenses for machine learning privacy forgo the provable guarantees of
differential privacy in the hope of achieving higher utility while resisting realistic adversaries …

Privacy side channels in machine learning systems

E Debenedetti, G Severi, N Carlini… - 33rd USENIX Security …, 2024 - usenix.org
Most current approaches for protecting privacy in machine learning (ML) assume that
models exist in a vacuum. Yet, in reality, these models are part of larger systems that include …

Unleashing the power of randomization in auditing differentially private ml

K Pillutla, G Andrew, P Kairouz… - Advances in …, 2024 - proceedings.neurips.cc
We present a rigorous methodology for auditing differentially private machine learning by
adding multiple carefully designed examples called canaries. We take a first principles …

One-shot empirical privacy estimation for federated learning

G Andrew, P Kairouz, S Oh, A Oprea… - arxiv preprint arxiv …, 2023 - arxiv.org
Privacy estimation techniques for differentially private (DP) algorithms are useful for
comparing against analytical bounds, or to empirically measure privacy loss in settings …

Preserving fairness and diagnostic accuracy in private large-scale AI models for medical imaging

S Tayebi Arasteh, A Ziller, C Kuhl, M Makowski… - Communications …, 2024 - nature.com
Background Artificial intelligence (AI) models are increasingly used in the medical domain.
However, as medical data is highly sensitive, special precautions to ensure its protection are …