Kernel mean embedding of distributions: A review and beyond

K Muandet, K Fukumizu… - … and Trends® in …, 2017 - nowpublishers.com
A Hilbert space embedding of a distribution—in short, a kernel mean embedding—has
recently emerged as a powerful tool for machine learning and statistical inference. The basic …

Gaussian processes and kernel methods: A review on connections and equivalences

M Kanagawa, P Hennig, D Sejdinovic… - arxiv preprint arxiv …, 2018 - arxiv.org
This paper is an attempt to bridge the conceptual gaps between researchers working on the
two widely used approaches based on positive definite kernels: Bayesian learning or …

[KNIHA][B] Computational topology for data analysis

TK Dey, Y Wang - 2022 - books.google.com
" In this chapter, we introduce some of the very basics that are used throughout the book.
First, we give the definition of a topological space and related notions of open and closed …

Learning operators with coupled attention

G Kissas, JH Seidman, LF Guilhoto… - Journal of Machine …, 2022 - jmlr.org
Supervised operator learning is an emerging machine learning paradigm with applications
to modeling the evolution of spatio-temporal dynamical systems and approximating general …

A kernel two-sample test

A Gretton, KM Borgwardt, MJ Rasch… - The Journal of Machine …, 2012 - dl.acm.org
We propose a framework for analyzing and comparing distributions, which we use to
construct statistical tests to determine if two samples are drawn from different distributions …

Domain adaptation with conditional transferable components

M Gong, K Zhang, T Liu, D Tao… - International …, 2016 - proceedings.mlr.press
Abstract Domain adaptation arises in supervised learning when the training (source domain)
and test (target domain) data have different distributions. Let X and Y denote the features …

Statistical inference on random dot product graphs: a survey

A Athreya, DE Fishkind, M Tang, CE Priebe… - Journal of Machine …, 2018 - jmlr.org
The random dot product graph (RDPG) is an independent-edge random graph that is
analytically tractable and, simultaneously, either encompasses or can successfully …

Domain adaptation under target and conditional shift

K Zhang, B Schölkopf, K Muandet… - … conference on machine …, 2013 - proceedings.mlr.press
Let X denote the feature and Y the target. We consider domain adaptation under three
possible scenarios:(1) the marginal P_Y changes, while the conditional P_X| Y stays the …

Equivalence of distance-based and RKHS-based statistics in hypothesis testing

D Sejdinovic, B Sriperumbudur, A Gretton… - The annals of …, 2013 - JSTOR
We provide a unifying framework linking two classes of statistics used in two-sample and
independence testing: on the one hand, the energy distances and distance covariances …

MMD-FUSE: Learning and combining kernels for two-sample testing without data splitting

F Biggs, A Schrab, A Gretton - Advances in Neural …, 2023 - proceedings.neurips.cc
We propose novel statistics which maximise the power of a two-sample test based on the
Maximum Mean Discrepancy (MMD), byadapting over the set of kernels used in defining it …