Generalization bounds: Perspectives from information theory and PAC-Bayes

F Hellström, G Durisi, B Guedj… - … and Trends® in …, 2025 - nowpublishers.com
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …

Information-theoretic generalization bounds for stochastic gradient descent

G Neu, GK Dziugaite, M Haghifam… - … on Learning Theory, 2021 - proceedings.mlr.press
We study the generalization properties of the popular stochastic optimization method known
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …

Information-theoretic generalization bounds for black-box learning algorithms

H Harutyunyan, M Raginsky… - Advances in Neural …, 2021 - proceedings.neurips.cc
We derive information-theoretic generalization bounds for supervised learning algorithms
based on the information contained in predictions rather than in the output of the training …

A new family of generalization bounds using samplewise evaluated CMI

F Hellström, G Durisi - Advances in Neural Information …, 2022 - proceedings.neurips.cc
We present a new family of information-theoretic generalization bounds, in which the
training loss and the population loss are compared through a jointly convex function. This …

Towards a unified information-theoretic framework for generalization

M Haghifam, GK Dziugaite… - Advances in Neural …, 2021 - proceedings.neurips.cc
In this work, we investigate the expressiveness of the" conditional mutual information"(CMI)
framework of Steinke and Zakynthinou (2020) and the prospect of using it to provide a …

Tighter expected generalization error bounds via Wasserstein distance

B Rodríguez Gálvez, G Bassi… - Advances in …, 2021 - proceedings.neurips.cc
This work presents several expected generalization error bounds based on the Wasserstein
distance. More specifically, it introduces full-dataset, single-letter, and random-subset …

Information complexity of stochastic convex optimization: Applications to generalization and memorization

I Attias, GK Dziugaite, M Haghifam, R Livni… - arxiv preprint arxiv …, 2024 - arxiv.org
In this work, we investigate the interplay between memorization and learning in the context
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …

Tighter information-theoretic generalization bounds from supersamples

Z Wang, Y Mao - arxiv preprint arxiv:2302.02432, 2023 - arxiv.org
In this work, we present a variety of novel information-theoretic generalization bounds for
learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)-the …

Limitations of information-theoretic generalization bounds for gradient descent methods in stochastic convex optimization

M Haghifam, B Rodríguez-Gálvez… - International …, 2023 - proceedings.mlr.press
To date, no “information-theoretic” frameworks for reasoning about generalization error have
been shown to establish minimax rates for gradient descent in the setting of stochastic …

Individually conditional individual mutual information bound on generalization error

R Zhou, C Tian, T Liu - IEEE Transactions on Information …, 2022 - ieeexplore.ieee.org
We propose an information-theoretic bound on the generalization error based on a
combination of the error decomposition technique of Bu et al. and the conditional mutual …