Inverse dynamics pretraining learns good representations for multitask imitation

D Brandfonbrener, O Nachum… - Advances in Neural …, 2023 - proceedings.neurips.cc
In recent years, domains such as natural language processing and image recognition have
popularized the paradigm of using large datasets to pretrain representations that can be …

Understanding and addressing the pitfalls of bisimulation-based representations in offline reinforcement learning

H Zang, X Li, L Zhang, Y Liu, B Sun… - Advances in …, 2023 - proceedings.neurips.cc
While bisimulation-based approaches hold promise for learning robust state representations
for Reinforcement Learning (RL) tasks, their efficacy in offline RL tasks has not been up to …

Learning latent dynamic robust representations for world models

R Sun, H Zang, X Li, R Islam - arxiv preprint arxiv:2405.06263, 2024 - arxiv.org
Visual Model-Based Reinforcement Learning (MBRL) promises to encapsulate agent's
knowledge about the underlying dynamics of the environment, enabling learning a world …

Disentangling policy from offline task representation learning via adversarial data augmentation

C Jia, F Zhang, YC Li, CX Gao, XH Liu, L Yuan… - arxiv preprint arxiv …, 2024 - arxiv.org
Offline meta-reinforcement learning (OMRL) proficiently allows an agent to tackle novel
tasks while solely relying on a static dataset. For precise and efficient task identification …

Investigating pre-training objectives for generalization in vision-based reinforcement learning

D Kim, H Lee, K Lee, D Hwang, J Choo - arxiv preprint arxiv:2406.06037, 2024 - arxiv.org
Recently, various pre-training methods have been introduced in vision-based
Reinforcement Learning (RL). However, their generalization ability remains unclear due to …

Making offline rl online: Collaborative world models for offline visual reinforcement learning

Q Wang, J Yang, Y Wang, X **… - Advances in Neural …, 2025 - proceedings.neurips.cc
Training offline RL models using visual inputs poses two significant challenges, ie, the
overfitting problem in representation learning and the overestimation bias for expected …

Principled offline RL in the presence of rich exogenous information

R Islam, M Tomar, A Lamb, Y Efroni, H Zang… - 2023 - openreview.net
Learning to control an agent from offline data collected in a rich pixel-based visual
observation space is vital for real-world applications of reinforcement learning (RL). A major …

Advantage-aware policy optimization for offline reinforcement learning

Y Qing, S Liu, J Cong, K Chen, Y Zhou, M Song - 2024 - openreview.net
Offline Reinforcement Learning (RL) endeavors to leverage offline datasets to craft effective
agent policy without online interaction, which imposes proper conservative constraints to …

基于表征学**的离线**化学**方法研究综述

王雪松, 王荣荣, 程玉虎 - 自动化学报, 2024 - aas.net.cn
**化学**(Reinforcement learning, RL) 通过智能体与环境在线交互来学**最优策略,
**年来已成为解决复杂环境下感知决策问题的重要手段. 然而, 在线收集数据的方式可能会引发 …

Bridging the Gap from Supervised Learning to Control

D Brandfonbrener - 2023 - search.proquest.com
The combination of deep learning and internet-scale data with supervised learning has led
to impressive progress in recent years. However, the potential of this progress has yet to be …