Is out-of-distribution detection learnable?

Z Fang, Y Li, J Lu, J Dong, B Han… - Advances in Neural …, 2022 - proceedings.neurips.cc
Supervised learning aims to train a classifier under the assumption that training and test
data are from the same distribution. To ease the above assumption, researchers have …

Dataset security for machine learning: Data poisoning, backdoor attacks, and defenses

M Goldblum, D Tsipras, C **e, X Chen… - … on Pattern Analysis …, 2022 - ieeexplore.ieee.org
As machine learning systems grow in scale, so do their training data requirements, forcing
practitioners to automate and outsource the curation of training data in order to achieve state …

Sample selection with uncertainty of losses for learning with noisy labels

X **a, T Liu, B Han, M Gong, J Yu, G Niu… - arxiv preprint arxiv …, 2021 - arxiv.org
In learning with noisy labels, the sample selection approach is very popular, which regards
small-loss data as correctly labeled during training. However, losses are generated on-the …

Robustness implies privacy in statistical estimation

SB Hopkins, G Kamath, M Majid… - Proceedings of the 55th …, 2023 - dl.acm.org
We study the relationship between adversarial robustness and differential privacy in high-
dimensional algorithmic statistics. We give the first black-box reduction from privacy to …

Differential privacy and robust statistics in high dimensions

X Liu, W Kong, S Oh - Conference on Learning Theory, 2022 - proceedings.mlr.press
We introduce a universal framework for characterizing the statistical efficiency of a statistical
estimation problem with differential privacy guarantees. Our framework, which we call High …

Efficient mean estimation with pure differential privacy via a sum-of-squares exponential mechanism

SB Hopkins, G Kamath, M Majid - Proceedings of the 54th Annual ACM …, 2022 - dl.acm.org
We give the first polynomial-time algorithm to estimate the mean of ad-variate probability
distribution with bounded covariance from Õ (d) independent samples subject to pure …

Privately estimating a gaussian: Efficient, robust, and optimal

D Alabi, PK Kothari, P Tankala, P Venkat… - Proceedings of the 55th …, 2023 - dl.acm.org
In this work, we give efficient algorithms for privately estimating a Gaussian distribution in
both pure and approximate differential privacy (DP) models with optimal dependence on the …

Robust and differentially private mean estimation

X Liu, W Kong, S Kakade, S Oh - Advances in neural …, 2021 - proceedings.neurips.cc
In statistical learning and analysis from shared data, which is increasingly widely adopted in
platforms such as federated learning and meta-learning, there are two major concerns …

Backdoor attacks against dataset distillation

Y Liu, Z Li, M Backes, Y Shen, Y Zhang - arxiv preprint arxiv:2301.01197, 2023 - arxiv.org
Dataset distillation has emerged as a prominent technique to improve data efficiency when
training machine learning models. It encapsulates the knowledge from a large dataset into a …

Robust sub-Gaussian estimation of a mean vector in nearly linear time

J Depersin, G Lecué - The Annals of Statistics, 2022 - projecteuclid.org
We construct an algorithm for estimating the mean of a heavy-tailed random variable when
given an adversarial corrupted sample of N independent observations. The only assumption …