To compress or not to compress—self-supervised learning and information theory: A review

R Shwartz Ziv, Y LeCun - Entropy, 2024 - mdpi.com
Deep neural networks excel in supervised learning tasks but are constrained by the need for
extensive labeled data. Self-supervised learning emerges as a promising alternative …

Privacy auditing with one (1) training run

T Steinke, M Nasr, M Jagielski - Advances in Neural …, 2023 - proceedings.neurips.cc
We propose a scheme for auditing differentially private machine learning systems with a
single training run. This exploits the parallelism of being able to add or remove multiple …

User-friendly introduction to PAC-Bayes bounds

P Alquier - Foundations and Trends® in Machine Learning, 2024 - nowpublishers.com
Aggregated predictors are obtained by making a set of basic predictors vote according to
some weights, that is, to some probability distribution. Randomized predictors are obtained …

Recent advances in deep learning theory

F He, D Tao - arxiv preprint arxiv:2012.10931, 2020 - arxiv.org
Deep learning is usually described as an experiment-driven field under continuous criticizes
of lacking theoretical foundations. This problem has been partially fixed by a large volume of …

Generalization bounds: Perspectives from information theory and PAC-Bayes

F Hellström, G Durisi, B Guedj… - … and Trends® in …, 2025 - nowpublishers.com
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …

IM-loss: information maximization loss for spiking neural networks

Y Guo, Y Chen, L Zhang, X Liu… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Spiking Neural Network (SNN), recognized as a type of biologically plausible
architecture, has recently drawn much research attention. It transmits information by $0/1 …

How does information bottleneck help deep learning?

K Kawaguchi, Z Deng, X Ji… - … Conference on Machine …, 2023 - proceedings.mlr.press
Numerous deep learning algorithms have been inspired by and understood via the notion of
information bottleneck, where unnecessary information is (often implicitly) minimized while …

Information-theoretic generalization bounds for stochastic gradient descent

G Neu, GK Dziugaite, M Haghifam… - … on Learning Theory, 2021 - proceedings.mlr.press
We study the generalization properties of the popular stochastic optimization method known
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …

Sharpened generalization bounds based on conditional mutual information and an application to noisy, iterative algorithms

M Haghifam, J Negrea, A Khisti… - Advances in …, 2020 - proceedings.neurips.cc
The information-theoretic framework of Russo and Zou (2016) and Xu and Raginsky (2017)
provides bounds on the generalization error of a learning algorithm in terms of the mutual …

An exact characterization of the generalization error for the Gibbs algorithm

G Aminian, Y Bu, L Toni… - Advances in Neural …, 2021 - proceedings.neurips.cc
Various approaches have been developed to upper bound the generalization error of a
supervised learning algorithm. However, existing bounds are often loose and lack of …