Efficient sim-to-real transfer of contact-rich manipulation skills with online admittance residual learning
Learning contact-rich manipulation skills is essential. Such skills require the robots to
interact with the environment with feasible manipulation trajectories and suitable compliance …
interact with the environment with feasible manipulation trajectories and suitable compliance …
Hierarchical planning through goal-conditioned offline reinforcement learning
Offline Reinforcement learning (RL) has shown potent in many safe-critical tasks in robotics
where exploration is risky and expensive. However, it still struggles to acquire skills in …
where exploration is risky and expensive. However, it still struggles to acquire skills in …
Design from policies: Conservative test-time adaptation for offline policy optimization
In this work, we decouple the iterative bi-level offline RL (value estimation and policy
extraction) from the offline training phase, forming a non-iterative bi-level paradigm and …
extraction) from the offline training phase, forming a non-iterative bi-level paradigm and …
When to trust your simulator: Dynamics-aware hybrid offline-and-online reinforcement learning
Learning effective reinforcement learning (RL) policies to solve real-world complex tasks
can be quite challenging without a high-fidelity simulation environment. In most cases, we …
can be quite challenging without a high-fidelity simulation environment. In most cases, we …
Diffusion policies for out-of-distribution generalization in offline reinforcement learning
Offline Reinforcement Learning (RL) methods leverage previous experiences to learn better
policies than the behavior policy used for data collection. However, they face challenges …
policies than the behavior policy used for data collection. However, they face challenges …
Guided online distillation: Promoting safe reinforcement learning by offline demonstration
Safe Reinforcement Learning (RL) aims to find a policy that achieves high rewards while
satisfying cost constraints. When learning from scratch, safe RL agents tend to be overly …
satisfying cost constraints. When learning from scratch, safe RL agents tend to be overly …
Adaptive prediction ensemble: Improving out-of-distribution generalization of motion forecasting
Deep learning-based trajectory prediction models for autonomous driving often struggle with
generalization to out-of-distribution (OOD) scenarios, sometimes performing worse than …
generalization to out-of-distribution (OOD) scenarios, sometimes performing worse than …
Odice: Revealing the mystery of distribution correction estimation via orthogonal-gradient update
In this study, we investigate the DIstribution Correction Estimation (DICE) methods, an
important line of work in offline reinforcement learning (RL) and imitation learning (IL). DICE …
important line of work in offline reinforcement learning (RL) and imitation learning (IL). DICE …
Domain: Mildly conservative model-based offline reinforcement learning
Model-based reinforcement learning (RL), which learns environment model from offline
dataset and generates more out-of-distribution model data, has become an effective …
dataset and generates more out-of-distribution model data, has become an effective …
[PDF][PDF] A Trajectory Perspective on the Role of Data Sampling Techniques in Offline Reinforcement Learning
In recent years, offline reinforcement learning (RL) algorithms have gained considerable
attention. However, the role of data sampling techniques in offline RL has been somewhat …
attention. However, the role of data sampling techniques in offline RL has been somewhat …