Maximum-entropy adversarial data augmentation for improved generalization and robustness

L Zhao, T Liu, X Peng… - Advances in Neural …, 2020 - proceedings.neurips.cc
Adversarial data augmentation has shown promise for training robust deep neural networks
against unforeseen data shifts or corruptions. However, it is difficult to define heuristics to …

Topics and techniques in distribution testing: A biased but representative sample

CL Canonne - Foundations and Trends® in Communications …, 2022 - nowpublishers.com
We focus on some specific problems in distribution testing, taking goodness-of-fit as a
running example. In particular, we do not aim to provide a comprehensive summary of all the …

Beyond normal: On the evaluation of mutual information estimators

P Czyż, F Grabowski, J Vogt… - Advances in Neural …, 2023 - proceedings.neurips.cc
Mutual information is a general statistical dependency measure which has found
applications in representation learning, causality, domain generalization and computational …

Hypothesis testing for high-dimensional multinomials: A selective review

S Balakrishnan, L Wasserman - 2018 - projecteuclid.org
The statistical analysis of discrete data has been the subject of extensive statistical research
dating back to the work of Pearson. In this survey we review some recently developed …

Estimating the number of species in microbial diversity studies

J Bunge, A Willis, F Walsh - Annual Review of Statistics and Its …, 2014 - annualreviews.org
For decades, statisticians have studied the species problem: how to estimate the total
number of species, observed plus unobserved, in a population. This problem dates at least …

Estimating mutual information for discrete-continuous mixtures

W Gao, S Kannan, S Oh… - Advances in neural …, 2017 - proceedings.neurips.cc
Estimation of mutual information from observed samples is a basic primitive in machine
learning, useful in several learning tasks including correlation mining, information …

Minimax estimation of functionals of discrete distributions

J Jiao, K Venkat, Y Han… - IEEE Transactions on …, 2015 - ieeexplore.ieee.org
We propose a general methodology for the construction and analysis of essentially minimax
estimators for a wide class of functionals of finite dimensional parameters, and elaborate on …

Do GANs learn the distribution? some theory and empirics

S Arora, A Risteski, Y Zhang - International conference on learning …, 2018 - openreview.net
Do GANS (Generative Adversarial Nets) actually learn the target distribution? The
foundational paper of Goodfellow et al.(2014) suggested they do, if they were given …

[SÁCH][B] Introduction to property testing

O Goldreich - 2017 - books.google.com
Property testing is concerned with the design of super-fast algorithms for the structural
analysis of large quantities of data. The aim is to unveil global features of the data, such as …

Minimax rates of entropy estimation on large alphabets via best polynomial approximation

Y Wu, P Yang - IEEE Transactions on Information Theory, 2016 - ieeexplore.ieee.org
Consider the problem of estimating the Shannon entropy of a distribution over elements from
independent samples. We show that the minimax mean-square error is within the universal …