Representation learning with multi-step inverse kinematics: An efficient and optimal approach to rich-observation rl

Z Mhammedi, DJ Foster… - … Conference on Machine …, 2023 - proceedings.mlr.press
We study the design of sample-efficient algorithms for reinforcement learning in the
presence of rich, high-dimensional observations, formalized via the Block MDP problem …

Inverse dynamics pretraining learns good representations for multitask imitation

D Brandfonbrener, O Nachum… - Advances in Neural …, 2023 - proceedings.neurips.cc
In recent years, domains such as natural language processing and image recognition have
popularized the paradigm of using large datasets to pretrain representations that can be …

Guide your agent with adaptive multimodal rewards

C Kim, Y Seo, H Liu, L Lee, J Shin… - Advances in Neural …, 2023 - proceedings.neurips.cc
Develo** an agent capable of adapting to unseen environments remains a difficult
challenge in imitation learning. This work presents Adaptive Return-conditioned Policy …

Ignorance is bliss: Robust control via information gating

M Tomar, R Islam, M Taylor… - Advances in Neural …, 2023 - proceedings.neurips.cc
Informational parsimony provides a useful inductive bias for learning representations that
achieve better generalization by being robust to noise and spurious correlations. We …

Learning latent dynamic robust representations for world models

R Sun, H Zang, X Li, R Islam - arxiv preprint arxiv:2405.06263, 2024 - arxiv.org
Visual Model-Based Reinforcement Learning (MBRL) promises to encapsulate agent's
knowledge about the underlying dynamics of the environment, enabling learning a world …

Investigating pre-training objectives for generalization in vision-based reinforcement learning

D Kim, H Lee, K Lee, D Hwang, J Choo - arxiv preprint arxiv:2406.06037, 2024 - arxiv.org
Recently, various pre-training methods have been introduced in vision-based
Reinforcement Learning (RL). However, their generalization ability remains unclear due to …

Masked and Inverse Dynamics Modeling for Data-Efficient Reinforcement Learning

YJ Lee, J Kim, YJ Park, M Kwak… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
In pixel-based deep reinforcement learning (DRL), learning representations of states that
change because of an agent's action or interaction with the environment poses a critical …

Rich-observation reinforcement learning with continuous latent dynamics

Y Song, L Wu, DJ Foster, A Krishnamurthy - arxiv preprint arxiv …, 2024 - arxiv.org
Sample-efficiency and reliability remain major bottlenecks toward wide adoption of
reinforcement learning algorithms in continuous settings with high-dimensional perceptual …

Video Occupancy Models

M Tomar, P Hansen-Estruch, P Bachman… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce a new family of video prediction models designed to support downstream
control tasks. We call these models Video Occupancy models (VOCs). VOCs operate in a …

PcLast: Discovering plannable continuous latent states

A Koul, S Sujit, S Chen, B Evans, L Wu, B Xu… - arxiv preprint arxiv …, 2023 - arxiv.org
Goal-conditioned planning benefits from learned low-dimensional representations of rich
observations. While compact latent representations typically learned from variational …