Convergence of entropy-regularized natural policy gradient with linear function approximation
Natural policy gradient (NPG) methods, equipped with function approximation and entropy
regularization, achieve impressive empirical success in reinforcement learning problems …
regularization, achieve impressive empirical success in reinforcement learning problems …
Information-theoretic generalization bounds for learning from quantum data
Learning tasks play an increasingly prominent role in quantum information and computation.
They range from fundamental problems such as state discrimination and metrology over the …
They range from fundamental problems such as state discrimination and metrology over the …
Information complexity of stochastic convex optimization: Applications to generalization and memorization
In this work, we investigate the interplay between memorization and learning in the context
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
of\emph {stochastic convex optimization}(SCO). We define memorization via the information …
Exactly tight information-theoretic generalization error bound for the quadratic gaussian problem
We provide a new information-theoretic generalization error bound that is exactly tight (ie,
matching even the constant) for the canonical quadratic Gaussian (location) problem. Most …
matching even the constant) for the canonical quadratic Gaussian (location) problem. Most …
An information-theoretic approach to generalization theory
We investigate the in-distribution generalization of machine learning algorithms. We depart
from traditional complexity-based approaches by analyzing information-theoretic bounds …
from traditional complexity-based approaches by analyzing information-theoretic bounds …
Empirical risk minimization with relative entropy regularization
The empirical risk minimization (ERM) problem with relative entropy regularization (ERM-
RER) is investigated under the assumption that the reference measure is a σ-finite measure …
RER) is investigated under the assumption that the reference measure is a σ-finite measure …
Tighter generalisation bounds via interpolation
This paper contains a recipe for deriving new PAC-Bayes generalisation bounds based on
the $(f,\Gamma) $-divergence, and, in addition, presents PAC-Bayes generalisation bounds …
the $(f,\Gamma) $-divergence, and, in addition, presents PAC-Bayes generalisation bounds …
Information-Theoretic Generalization Bounds for Transductive Learning and its Applications
H Tang, Y Liu - arxiv preprint arxiv:2311.04561, 2023 - arxiv.org
In this paper, we develop data-dependent and algorithm-dependent generalization bounds
for transductive learning algorithms in the context of information theory for the first time. We …
for transductive learning algorithms in the context of information theory for the first time. We …
Generalization and informativeness of conformal prediction
The safe integration of machine learning modules in decision-making processes hinges on
their ability to quantify uncertainty. A popular technique to achieve this goal is conformal …
their ability to quantify uncertainty. A popular technique to achieve this goal is conformal …
Empirical risk minimization with relative entropy regularization type-II
The effect of the relative entropy asymmetry is analyzed in the empirical risk minimization
with relative entropy regularization (ERM-RER) problem. A novel regularization is …
with relative entropy regularization (ERM-RER) problem. A novel regularization is …