Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Generalization bounds: Perspectives from information theory and PAC-Bayes
A fundamental question in theoretical machine learning is generalization. Over the past
decades, the PAC-Bayesian approach has been established as a flexible framework to …
decades, the PAC-Bayesian approach has been established as a flexible framework to …
Information-theoretic generalization bounds for stochastic gradient descent
We study the generalization properties of the popular stochastic optimization method known
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
as stochastic gradient descent (SGD) for optimizing general non-convex loss functions. Our …
Information-theoretic generalization bounds for black-box learning algorithms
We derive information-theoretic generalization bounds for supervised learning algorithms
based on the information contained in predictions rather than in the output of the training …
based on the information contained in predictions rather than in the output of the training …
A new family of generalization bounds using samplewise evaluated CMI
We present a new family of information-theoretic generalization bounds, in which the
training loss and the population loss are compared through a jointly convex function. This …
training loss and the population loss are compared through a jointly convex function. This …
Towards a unified information-theoretic framework for generalization
In this work, we investigate the expressiveness of the" conditional mutual information"(CMI)
framework of Steinke and Zakynthinou (2020) and the prospect of using it to provide a …
framework of Steinke and Zakynthinou (2020) and the prospect of using it to provide a …
Tighter expected generalization error bounds via Wasserstein distance
This work presents several expected generalization error bounds based on the Wasserstein
distance. More specifically, it introduces full-dataset, single-letter, and random-subset …
distance. More specifically, it introduces full-dataset, single-letter, and random-subset …
Information complexity of stochastic convex optimization: Applications to generalization and memorization
In this work, we investigate the interplay between memorization and learning in the context
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
Tighter information-theoretic generalization bounds from supersamples
In this work, we present a variety of novel information-theoretic generalization bounds for
learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)-the …
learning algorithms, from the supersample setting of Steinke & Zakynthinou (2020)-the …
Limitations of information-theoretic generalization bounds for gradient descent methods in stochastic convex optimization
To date, no “information-theoretic” frameworks for reasoning about generalization error have
been shown to establish minimax rates for gradient descent in the setting of stochastic …
been shown to establish minimax rates for gradient descent in the setting of stochastic …
Individually conditional individual mutual information bound on generalization error
We propose an information-theoretic bound on the generalization error based on a
combination of the error decomposition technique of Bu et al. and the conditional mutual …
combination of the error decomposition technique of Bu et al. and the conditional mutual …