Guiding pretraining in reinforcement learning with large language models

Y Du, O Watkins, Z Wang, C Colas… - International …, 2023 - proceedings.mlr.press
Reinforcement learning algorithms typically struggle in the absence of a dense, well-shaped
reward function. Intrinsically motivated exploration methods address this limitation by …

A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning

EF Morales, R Murrieta-Cid, I Becerra… - Intelligent Service …, 2021 - Springer
This article is about deep learning (DL) and deep reinforcement learning (DRL) works
applied to robotics. Both tools have been shown to be successful in delivering data-driven …

Mastering visual continuous control: Improved data-augmented reinforcement learning

D Yarats, R Fergus, A Lazaric, L Pinto - arxiv preprint arxiv:2107.09645, 2021 - arxiv.org
We present DrQ-v2, a model-free reinforcement learning (RL) algorithm for visual
continuous control. DrQ-v2 builds on DrQ, an off-policy actor-critic approach that uses data …

An information-theoretic perspective on intrinsic motivation in reinforcement learning: A survey

A Aubret, L Matignon, S Hassas - Entropy, 2023 - mdpi.com
The reinforcement learning (RL) research area is very active, with an important number of
new contributions, especially considering the emergent field of deep RL (DRL). However, a …

Masked world models for visual control

Y Seo, D Hafner, H Liu, F Liu, S James… - … on Robot Learning, 2023 - proceedings.mlr.press
Visual model-based reinforcement learning (RL) has the potential to enable sample-efficient
robot learning from visual observations. Yet the current approaches typically train a single …

Reinforcement learning with action-free pre-training from videos

Y Seo, K Lee, SL James… - … Conference on Machine …, 2022 - proceedings.mlr.press
Recent unsupervised pre-training methods have shown to be effective on language and
vision domains by learning useful representations for multiple downstream tasks. In this …

Behavior from the void: Unsupervised active pre-training

H Liu, P Abbeel - Advances in Neural Information …, 2021 - proceedings.neurips.cc
We introduce a new unsupervised pre-training method for reinforcement learning called
APT, which stands for Active Pre-Training. APT learns behaviors and representations by …

Aps: Active pretraining with successor features

H Liu, P Abbeel - International Conference on Machine …, 2021 - proceedings.mlr.press
We introduce a new unsupervised pretraining objective for reinforcement learning. During
the unsupervised reward-free pretraining phase, the agent maximizes mutual information …

Pretraining representations for data-efficient reinforcement learning

M Schwarzer, N Rajkumar… - Advances in …, 2021 - proceedings.neurips.cc
Data efficiency is a key challenge for deep reinforcement learning. We address this problem
by using unlabeled data to pretrain an encoder which is then finetuned on a small amount of …

Urlb: Unsupervised reinforcement learning benchmark

M Laskin, D Yarats, H Liu, K Lee, A Zhan, K Lu… - arxiv preprint arxiv …, 2021 - arxiv.org
Deep Reinforcement Learning (RL) has emerged as a powerful paradigm to solve a range
of complex yet specific control tasks. Yet training generalist agents that can quickly adapt to …