Representation learning with multi-step inverse kinematics: An efficient and optimal approach to rich-observation rl
We study the design of sample-efficient algorithms for reinforcement learning in the
presence of rich, high-dimensional observations, formalized via the Block MDP problem …
presence of rich, high-dimensional observations, formalized via the Block MDP problem …
Inverse dynamics pretraining learns good representations for multitask imitation
In recent years, domains such as natural language processing and image recognition have
popularized the paradigm of using large datasets to pretrain representations that can be …
popularized the paradigm of using large datasets to pretrain representations that can be …
Ignorance is bliss: Robust control via information gating
Informational parsimony provides a useful inductive bias for learning representations that
achieve better generalization by being robust to noise and spurious correlations. We …
achieve better generalization by being robust to noise and spurious correlations. We …
Guide your agent with adaptive multimodal rewards
Develo** an agent capable of adapting to unseen environments remains a difficult
challenge in imitation learning. This work presents Adaptive Return-conditioned Policy …
challenge in imitation learning. This work presents Adaptive Return-conditioned Policy …
Masked and Inverse Dynamics Modeling for Data-Efficient Reinforcement Learning
In pixel-based deep reinforcement learning (DRL), learning representations of states that
change because of an agent's action or interaction with the environment poses a critical …
change because of an agent's action or interaction with the environment poses a critical …
Video Occupancy Models
We introduce a new family of video prediction models designed to support downstream
control tasks. We call these models Video Occupancy models (VOCs). VOCs operate in a …
control tasks. We call these models Video Occupancy models (VOCs). VOCs operate in a …
PcLast: Discovering Plannable Continuous Latent States
Goal-conditioned planning benefits from learned low-dimensional representations of rich,
high-dimensional observations. While compact latent representations, typically learned from …
high-dimensional observations. While compact latent representations, typically learned from …
Investigating Pre-Training Objectives for Generalization in Vision-Based Reinforcement Learning
Recently, various pre-training methods have been introduced in vision-based
Reinforcement Learning (RL). However, their generalization ability remains unclear due to …
Reinforcement Learning (RL). However, their generalization ability remains unclear due to …
Rich-Observation Reinforcement Learning with Continuous Latent Dynamics
Sample-efficiency and reliability remain major bottlenecks toward wide adoption of
reinforcement learning algorithms in continuous settings with high-dimensional perceptual …
reinforcement learning algorithms in continuous settings with high-dimensional perceptual …
Towards Principled Representation Learning from Videos for Reinforcement Learning
We study pre-training representations for decision-making using video data, which is
abundantly available for tasks such as game agents and software testing. Even though …
abundantly available for tasks such as game agents and software testing. Even though …