Kernel mean embedding of distributions: A review and beyond

K Muandet, K Fukumizu… - … and Trends® in …, 2017 - nowpublishers.com
A Hilbert space embedding of a distribution—in short, a kernel mean embedding—has
recently emerged as a powerful tool for machine learning and statistical inference. The basic …

Examples are not enough, learn to criticize! criticism for interpretability

B Kim, R Khanna, OO Koyejo - Advances in neural …, 2016 - proceedings.neurips.cc
Example-based explanations are widely used in the effort to improve the interpretability of
highly complex distributions. However, prototypes alone are rarely sufficient to represent the …

Near-linear time approximation algorithms for optimal transport via Sinkhorn iteration

J Altschuler, J Niles-Weed… - Advances in neural …, 2017 - proceedings.neurips.cc
Computing optimal transport distances such as the earth mover's distance is a fundamental
problem in machine learning, statistics, and computer vision. Despite the recent introduction …

Revisiting classifier two-sample tests

D Lopez-Paz, M Oquab - arxiv preprint arxiv:1610.06545, 2016 - arxiv.org
The goal of two-sample tests is to assess whether two samples, $ S_P\sim P^ n $ and $
S_Q\sim Q^ m $, are drawn from the same distribution. Perhaps intriguingly, one relatively …

A kernel test of goodness of fit

K Chwialkowski, H Strathmann… - … conference on machine …, 2016 - proceedings.mlr.press
We propose a nonparametric statistical test for goodness-of-fit: given a set of samples, the
test determines how likely it is that these were generated from a target density function. The …

Making tree ensembles interpretable: A bayesian model selection approach

S Hara, K Hayashi - International conference on artificial …, 2018 - proceedings.mlr.press
Tree ensembles, such as random forests, are renowned for their high prediction
performance. However, their interpretability is critically limited due to the enormous …

Generative models and model criticism via optimized maximum mean discrepancy

DJ Sutherland, HY Tung, H Strathmann, S De… - arxiv preprint arxiv …, 2016 - arxiv.org
We propose a method to optimize the representation and distinguishability of samples from
two probability distributions, by maximizing the estimated power of a statistical test based on …

One-network adversarial fairness

T Adel, I Valera, Z Ghahramani, A Weller - Proceedings of the AAAI …, 2019 - ojs.aaai.org
There is currently a great expansion of the impact of machine learning algorithms on our
lives, prompting the need for objectives other than pure performance, including fairness …

Fast two-sample testing with analytic representations of probability measures

KP Chwialkowski, A Ramdas… - Advances in Neural …, 2015 - proceedings.neurips.cc
We propose a class of nonparametric two-sample tests with a cost linear in the sample size.
Two tests are given, both based on an ensemble of distances between analytic functions …

Interpretable distribution features with maximum testing power

W Jitkrittum, Z Szabó… - Advances in Neural …, 2016 - proceedings.neurips.cc
Two semimetrics on probability distributions are proposed, given as the sum of differences of
expectations of analytic functions evaluated at spatial or frequency locations (ie, features) …