The role of coverage in online reinforcement learning

T **e, DJ Foster, Y Bai, N Jiang, SM Kakade - arxiv preprint arxiv …, 2022 - arxiv.org
Coverage conditions--which assert that the data logging distribution adequately covers the
state space--play a fundamental role in determining the sample complexity of offline …

Representation learning with multi-step inverse kinematics: An efficient and optimal approach to rich-observation rl

Z Mhammedi, DJ Foster… - … Conference on Machine …, 2023 - proceedings.mlr.press
We study the design of sample-efficient algorithms for reinforcement learning in the
presence of rich, high-dimensional observations, formalized via the Block MDP problem …

Smart: Self-supervised multi-task pretraining with control transformers

Y Sun, S Ma, R Madaan, R Bonatti, F Huang… - arxiv preprint arxiv …, 2023 - arxiv.org
Self-supervised pretraining has been extensively studied in language and vision domains,
where a unified model can be easily adapted to various downstream tasks by pretraining …

Inverse dynamics pretraining learns good representations for multitask imitation

D Brandfonbrener, O Nachum… - Advances in Neural …, 2023 - proceedings.neurips.cc
In recent years, domains such as natural language processing and image recognition have
popularized the paradigm of using large datasets to pretrain representations that can be …

Ignorance is bliss: Robust control via information gating

M Tomar, R Islam, M Taylor… - Advances in Neural …, 2023 - proceedings.neurips.cc
Informational parsimony provides a useful inductive bias for learning representations that
achieve better generalization by being robust to noise and spurious correlations. We …

A unified view on solving objective mismatch in model-based reinforcement learning

R Wei, N Lambert, A McDonald, A Garcia… - arxiv preprint arxiv …, 2023 - arxiv.org
Model-based Reinforcement Learning (MBRL) aims to make agents more sample-efficient,
adaptive, and explainable by learning an explicit model of the environment. While the …

Discrete factorial representations as an abstraction for goal conditioned reinforcement learning

R Islam, H Zang, A Goyal, A Lamb… - arxiv preprint arxiv …, 2022 - arxiv.org
Goal-conditioned reinforcement learning (RL) is a promising direction for training agents that
are capable of solving multiple tasks and reach a diverse set of objectives. How to\textit …

Agent-controller representations: Principled offline rl with rich exogenous information

R Islam, M Tomar, A Lamb, Y Efroni, H Zang… - arxiv preprint arxiv …, 2022 - arxiv.org
Learning to control an agent from data collected offline in a rich pixel-based visual
observation space is vital for real-world applications of reinforcement learning (RL). A major …

Discrete compositional representations as an abstraction for goal conditioned reinforcement learning

R Islam, H Zang, A Goyal, AM Lamb… - Advances in …, 2022 - proceedings.neurips.cc
Goal-conditioned reinforcement learning (RL) is a promising direction for training agents that
are capable of solving multiple tasks and reach a diverse set of objectives. How to\textit …

Rich-observation reinforcement learning with continuous latent dynamics

Y Song, L Wu, DJ Foster, A Krishnamurthy - arxiv preprint arxiv …, 2024 - arxiv.org
Sample-efficiency and reliability remain major bottlenecks toward wide adoption of
reinforcement learning algorithms in continuous settings with high-dimensional perceptual …