Generalization bounds: Perspectives from information theory and PAC-Bayes
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …
decades, the PAC-Bayesian approach has been established as a flexible framework to …
Tighter information-theoretic generalization bounds from supersamples
In this work, we present a variety of novel information-theoretic generalization bounds for
learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)-the …
learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)-the …
Information complexity of stochastic convex optimization: Applications to generalization and memorization
In this work, we investigate the interplay between memorization and learning in the context
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
Bayes meets bernstein at the meta level: an analysis of fast rates in meta-learning with pac-bayes
Bernstein's condition is a key assumption that guarantees fast rates in machine learning. For
example, the Gibbs algorithm with prior $\pi $ has an excess risk in $ O (d_ {\pi}/n) $, as …
example, the Gibbs algorithm with prior $\pi $ has an excess risk in $ O (d_ {\pi}/n) $, as …
Exactly tight information-theoretic generalization error bound for the quadratic gaussian problem
We provide a new information-theoretic generalization error bound that is exactly tight (ie,
matching even the constant) for the canonical quadratic Gaussian (location) problem. Most …
matching even the constant) for the canonical quadratic Gaussian (location) problem. Most …
Learning an explicit hyper-parameter prediction function conditioned on tasks
Meta learning has attracted much attention recently in machine learning community.
Contrary to conventional machine learning aiming to learn inherent prediction rules to …
Contrary to conventional machine learning aiming to learn inherent prediction rules to …
On the generalization error of meta learning for the Gibbs algorithm
We analyze the generalization ability of joint-training meta learning algorithms via the Gibbs
algorithm. Our exact characterization of the expected meta generalization error for the meta …
algorithm. Our exact characterization of the expected meta generalization error for the meta …
More Flexible PAC-Bayesian Meta-Learning by Learning Learning Algorithms
We introduce a new framework for studying meta-learning methods using PAC-Bayesian
theory. Its main advantage over previous work is that it allows for more flexibility in how the …
theory. Its main advantage over previous work is that it allows for more flexibility in how the …
Error Bounds of Supervised Classification from Information-Theoretic Perspective
B Qi, W Gong, L Li - arxiv preprint arxiv:2406.04567, 2024 - arxiv.org
There remains a list of unanswered research questions on deep learning (DL), including the
remarkable generalization power of overparametrized neural networks, the efficient …
remarkable generalization power of overparametrized neural networks, the efficient …
Towards Sharper Information-theoretic Generalization Bounds for Meta-Learning
In recent years, information-theoretic generalization bounds have emerged as a promising
approach for analyzing the generalization capabilities of meta-learning algorithms …
approach for analyzing the generalization capabilities of meta-learning algorithms …