Large sequence models for sequential decision-making: a survey

M Wen, R Lin, H Wang, Y Yang, Y Wen, L Mai… - Frontiers of Computer …, 2023 - Springer
Transformer architectures have facilitated the development of large-scale and general-
purpose sequence models for prediction tasks in natural language processing and computer …

Large language models as general pattern machines

S Mirchandani, F **a, P Florence, B Ichter… - arxiv preprint arxiv …, 2023 - arxiv.org
We observe that pre-trained large language models (LLMs) are capable of autoregressively
completing complex token sequences--from arbitrary ones procedurally generated by …

Supervised pretraining can learn in-context reinforcement learning

J Lee, A **e, A Pacchiano, Y Chandak… - Advances in …, 2024 - proceedings.neurips.cc
Large transformer models trained on diverse datasets have shown a remarkable ability to
learn in-context, achieving high few-shot performance on tasks they were not explicitly …

On Transforming Reinforcement Learning With Transformers: The Development Trajectory

S Hu, L Shen, Y Zhang, Y Chen… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Transformers, originally devised for natural language processing (NLP), have also produced
significant successes in computer vision (CV). Due to their strong expression power …

Diffusion model is an effective planner and data synthesizer for multi-task reinforcement learning

H He, C Bai, K Xu, Z Yang, W Zhang… - Advances in neural …, 2024 - proceedings.neurips.cc
Diffusion models have demonstrated highly-expressive generative capabilities in vision and
NLP. Recent studies in reinforcement learning (RL) have shown that diffusion models are …

In-context reinforcement learning with algorithm distillation

M Laskin, L Wang, J Oh, E Parisotto, S Spencer… - arxiv preprint arxiv …, 2022 - arxiv.org
We propose Algorithm Distillation (AD), a method for distilling reinforcement learning (RL)
algorithms into neural networks by modeling their training histories with a causal sequence …

Elastic decision transformer

YH Wu, X Wang, M Hamaya - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract This paper introduces Elastic Decision Transformer (EDT), a significant
advancement over the existing Decision Transformer (DT) and its variants. Although DT …

Constrained decision transformer for offline safe reinforcement learning

Z Liu, Z Guo, Y Yao, Z Cen, W Yu… - International …, 2023 - proceedings.mlr.press
Safe reinforcement learning (RL) trains a constraint satisfaction policy by interacting with the
environment. We aim to tackle a more challenging problem: learning a safe policy from an …

Skill transformer: A monolithic policy for mobile manipulation

X Huang, D Batra, A Rai, A Szot - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract We present Skill Transformer, an approach for solving long-horizon robotic tasks by
combining conditional sequence modeling and skill modularity. Conditioned on egocentric …

Ceil: Generalized contextual imitation learning

J Liu, L He, Y Kang, Z Zhuang… - Advances in Neural …, 2023 - proceedings.neurips.cc
In this paper, we present ContExtual Imitation Learning (CEIL), a general and broadly
applicable algorithm for imitation learning (IL). Inspired by the formulation of hindsight …