Look before you leap: Unveiling the power of gpt-4v in robotic vision-language planning

Y Hu, F Lin, T Zhang, L Yi, Y Gao - arxiv preprint arxiv:2311.17842, 2023 - arxiv.org
In this study, we are interested in imbuing robots with the capability of physically-grounded
task planning. Recent advancements have shown that large language models (LLMs) …

Bridgedata v2: A dataset for robot learning at scale

HR Walke, K Black, TZ Zhao, Q Vuong… - … on Robot Learning, 2023 - proceedings.mlr.press
We introduce BridgeData V2, a large and diverse dataset of robotic manipulation behaviors
designed to facilitate research in scalable robot learning. BridgeData V2 contains 53,896 …

Vip: Towards universal visual reward and representation via value-implicit pre-training

YJ Ma, S Sodhani, D Jayaraman, O Bastani… - arxiv preprint arxiv …, 2022 - arxiv.org
Reward and representation learning are two long-standing challenges for learning an
expanding set of robot manipulation skills from sensory observations. Given the inherent …

Td-mpc2: Scalable, robust world models for continuous control

N Hansen, H Su, X Wang - arxiv preprint arxiv:2310.16828, 2023 - arxiv.org
TD-MPC is a model-based reinforcement learning (RL) algorithm that performs local
trajectory optimization in the latent space of a learned implicit (decoder-free) world model. In …

Goal-conditioned imitation learning using score-based diffusion policies

M Reuss, M Li, X Jia, R Lioutikov - arxiv preprint arxiv:2304.02532, 2023 - arxiv.org
We propose a new policy representation based on score-based diffusion models (SDMs).
We apply our new policy representation in the domain of Goal-Conditioned Imitation …

Hiql: Offline goal-conditioned rl with latent states as actions

S Park, D Ghosh, B Eysenbach… - Advances in Neural …, 2023 - proceedings.neurips.cc
Unsupervised pre-training has recently become the bedrock for computer vision and natural
language processing. In reinforcement learning (RL), goal-conditioned RL can potentially …

Factorized contrastive learning: Going beyond multi-view redundancy

PP Liang, Z Deng, MQ Ma, JY Zou… - Advances in …, 2023 - proceedings.neurips.cc
In a wide range of multimodal tasks, contrastive learning has become a particularly
appealing approach since it can successfully learn representations from abundant …

: Temporal Latent Action-Driven Contrastive Loss for Visual Reinforcement Learning

R Zheng, X Wang, Y Sun, S Ma… - Advances in …, 2023 - proceedings.neurips.cc
Despite recent progress in reinforcement learning (RL) from raw pixel data, sample
inefficiency continues to present a substantial obstacle. Prior works have attempted to …

Inference via interpolation: Contrastive representations provably enable planning and inference

B Eysenbach, V Myers… - Advances in Neural …, 2025 - proceedings.neurips.cc
Given time series data, how can we answer questions like what will happen in the
future?''and how did we get here?''These sorts of probabilistic inference questions are …

Reinforcement learning from passive data via latent intentions

D Ghosh, CA Bhateja, S Levine - … Conference on Machine …, 2023 - proceedings.mlr.press
Passive observational data, such as human videos, is abundant and rich in information, yet
remains largely untapped by current RL methods. Perhaps surprisingly, we show that …