Explainable reinforcement learning: A survey and comparative review

S Milani, N Topin, M Veloso, F Fang - ACM Computing Surveys, 2024 - dl.acm.org
Explainable reinforcement learning (XRL) is an emerging subfield of explainable machine
learning that has attracted considerable attention in recent years. The goal of XRL is to …

Ring attention with blockwise transformers for near-infinite context

H Liu, M Zaharia, P Abbeel - arxiv preprint arxiv:2310.01889, 2023 - arxiv.org
Transformers have emerged as the architecture of choice for many state-of-the-art AI
models, showcasing exceptional performance across a wide range of AI applications …

A comprehensive survey of data augmentation in visual reinforcement learning

G Ma, Z Wang, Z Yuan, X Wang, B Yuan… - arxiv preprint arxiv …, 2022 - arxiv.org
Visual reinforcement learning (RL), which makes decisions directly from high-dimensional
visual inputs, has demonstrated significant potential in various domains. However …

Reinforcement learning with action-free pre-training from videos

Y Seo, K Lee, SL James… - … Conference on Machine …, 2022 - proceedings.mlr.press
Recent unsupervised pre-training methods have shown to be effective on language and
vision domains by learning useful representations for multiple downstream tasks. In this …

Masked trajectory models for prediction, representation, and control

P Wu, A Majumdar, K Stone, Y Lin… - International …, 2023 - proceedings.mlr.press
Abstract We introduce Masked Trajectory Models (MTM) as a generic abstraction for
sequential decision making. MTM takes a trajectory, such as a state-action sequence, and …

B-pref: Benchmarking preference-based reinforcement learning

K Lee, L Smith, A Dragan, P Abbeel - arxiv preprint arxiv:2111.03026, 2021 - arxiv.org
Reinforcement learning (RL) requires access to a reward function that incentivizes the right
behavior, but these are notoriously hard to specify for complex tasks. Preference-based RL …

Blockwise parallel transformers for large context models

H Liu, P Abbeel - Advances in neural information …, 2023 - proceedings.neurips.cc
Transformers have emerged as the cornerstone of state-of-the-art natural language
processing models, showcasing exceptional performance across a wide range of AI …

Pre-training contextualized world models with in-the-wild videos for reinforcement learning

J Wu, H Ma, C Deng, M Long - Advances in Neural …, 2023 - proceedings.neurips.cc
Unsupervised pre-training methods utilizing large and diverse datasets have achieved
tremendous success across a range of domains. Recent work has investigated such …

Emergent agentic transformer from chain of hindsight experience

H Liu, P Abbeel - International Conference on Machine …, 2023 - proceedings.mlr.press
Large transformer models powered by diverse data and model scale have dominated
natural language modeling and computer vision and pushed the frontier of multiple AI areas …

Controllability-aware unsupervised skill discovery

S Park, K Lee, Y Lee, P Abbeel - arxiv preprint arxiv:2302.05103, 2023 - arxiv.org
One of the key capabilities of intelligent agents is the ability to discover useful skills without
external supervision. However, the current unsupervised skill discovery methods are often …