Bellman eluder dimension: New rich classes of rl problems, and sample-efficient algorithms

C **, Q Liu, S Miryoosefi - Advances in neural information …, 2021 - proceedings.neurips.cc
Finding the minimal structural assumptions that empower sample-efficient learning is one of
the most important research directions in Reinforcement Learning (RL). This paper …

Nearly minimax optimal reinforcement learning for linear mixture markov decision processes

D Zhou, Q Gu, C Szepesvari - Conference on Learning …, 2021 - proceedings.mlr.press
We study reinforcement learning (RL) with linear function approximation where the
underlying transition probability kernel of the Markov decision process (MDP) is a linear …

Optimality and approximation with policy gradient methods in markov decision processes

A Agarwal, SM Kakade, JD Lee… - … on Learning Theory, 2020 - proceedings.mlr.press
Policy gradient (PG) methods are among the most effective methods in challenging
reinforcement learning problems with large state and/or action spaces. However, little is …

Information-theoretic considerations in batch reinforcement learning

J Chen, N Jiang - International Conference on Machine …, 2019 - proceedings.mlr.press
Value-function approximation methods that operate in batch mode have foundational
importance to reinforcement learning (RL). Finite sample guarantees for these methods …

Representation learning for online and offline rl in low-rank mdps

M Uehara, X Zhang, W Sun - arxiv preprint arxiv:2110.04652, 2021 - arxiv.org
This work studies the question of Representation Learning in RL: how can we learn a
compact low-dimensional representation such that on top of the representation we can …

Nearly minimax optimal reinforcement learning for linear markov decision processes

J He, H Zhao, D Zhou, Q Gu - International Conference on …, 2023 - proceedings.mlr.press
We study reinforcement learning (RL) with linear function approximation. For episodic time-
inhomogeneous linear Markov decision processes (linear MDPs) whose transition …

Learning near optimal policies with low inherent bellman error

A Zanette, A Lazaric, M Kochenderfer… - International …, 2020 - proceedings.mlr.press
We study the exploration problem with approximate linear action-value functions in episodic
reinforcement learning under the notion of low inherent Bellman error, a condition normally …

Is a good representation sufficient for sample efficient reinforcement learning?

SS Du, SM Kakade, R Wang, LF Yang - arxiv preprint arxiv:1910.03016, 2019 - arxiv.org
Modern deep learning methods provide effective means to learn good representations.
However, is a good representation itself sufficient for sample efficient reinforcement …

Provably efficient rl with rich observations via latent state decoding

S Du, A Krishnamurthy, N Jiang… - International …, 2019 - proceedings.mlr.press
We study the exploration problem in episodic MDPs with rich observations generated from a
small number of latent states. Under certain identifiability assumptions, we demonstrate how …

Model-based rl in contextual decision processes: Pac bounds and exponential improvements over model-free approaches

W Sun, N Jiang, A Krishnamurthy… - … on learning theory, 2019 - proceedings.mlr.press
We study the sample complexity of model-based reinforcement learning (henceforth RL) in
general contextual decision processes that require strategic exploration to find a near …