To compress or not to compress—self-supervised learning and information theory: A review
Deep neural networks excel in supervised learning tasks but are constrained by the need for
extensive labeled data. Self-supervised learning emerges as a promising alternative …
extensive labeled data. Self-supervised learning emerges as a promising alternative …
Privacy auditing with one (1) training run
We propose a scheme for auditing differentially private machine learning systems with a
single training run. This exploits the parallelism of being able to add or remove multiple …
single training run. This exploits the parallelism of being able to add or remove multiple …
User-friendly introduction to PAC-Bayes bounds
P Alquier - Foundations and Trends® in Machine Learning, 2024 - nowpublishers.com
Aggregated predictors are obtained by making a set of basic predictors vote according to
some weights, that is, to some probability distribution. Randomized predictors are obtained …
some weights, that is, to some probability distribution. Randomized predictors are obtained …
Recent advances in deep learning theory
Deep learning is usually described as an experiment-driven field under continuous criticizes
of lacking theoretical foundations. This problem has been partially fixed by a large volume of …
of lacking theoretical foundations. This problem has been partially fixed by a large volume of …
Generalization bounds: Perspectives from information theory and PAC-Bayes
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …
decades, the PAC-Bayesian approach has been established as a flexible framework to …
IM-loss: information maximization loss for spiking neural networks
Y Guo, Y Chen, L Zhang, X Liu… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Spiking Neural Network (SNN), recognized as a type of biologically plausible
architecture, has recently drawn much research attention. It transmits information by $0/1 …
architecture, has recently drawn much research attention. It transmits information by $0/1 …
How does information bottleneck help deep learning?
Numerous deep learning algorithms have been inspired by and understood via the notion of
information bottleneck, where unnecessary information is (often implicitly) minimized while …
information bottleneck, where unnecessary information is (often implicitly) minimized while …
Information-theoretic generalization bounds for stochastic gradient descent
We study the generalization properties of the popular stochastic optimization method known
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
Sharpened generalization bounds based on conditional mutual information and an application to noisy, iterative algorithms
The information-theoretic framework of Russo and Zou (2016) and Xu and Raginsky (2017)
provides bounds on the generalization error of a learning algorithm in terms of the mutual …
provides bounds on the generalization error of a learning algorithm in terms of the mutual …
An exact characterization of the generalization error for the Gibbs algorithm
Various approaches have been developed to upper bound the generalization error of a
supervised learning algorithm. However, existing bounds are often loose and lack of …
supervised learning algorithm. However, existing bounds are often loose and lack of …