Efficient sim-to-real transfer of contact-rich manipulation skills with online admittance residual learning
Learning contact-rich manipulation skills is essential. Such skills require the robots to
interact with the environment with feasible manipulation trajectories and suitable compliance …
interact with the environment with feasible manipulation trajectories and suitable compliance …
Hierarchical planning through goal-conditioned offline reinforcement learning
Offline Reinforcement learning (RL) has shown potent in many safe-critical tasks in robotics
where exploration is risky and expensive. However, it still struggles to acquire skills in …
where exploration is risky and expensive. However, it still struggles to acquire skills in …
Design from policies: Conservative test-time adaptation for offline policy optimization
In this work, we decouple the iterative bi-level offline RL (value estimation and policy
extraction) from the offline training phase, forming a non-iterative bi-level paradigm and …
extraction) from the offline training phase, forming a non-iterative bi-level paradigm and …
When to trust your simulator: Dynamics-aware hybrid offline-and-online reinforcement learning
Learning effective reinforcement learning (RL) policies to solve real-world complex tasks
can be quite challenging without a high-fidelity simulation environment. In most cases, we …
can be quite challenging without a high-fidelity simulation environment. In most cases, we …
Diffusion policies for out-of-distribution generalization in offline reinforcement learning
Offline Reinforcement Learning (RL) methods leverage previous experiences to learn better
policies than the behavior policy used for data collection. However, they face challenges …
policies than the behavior policy used for data collection. However, they face challenges …
Guided online distillation: Promoting safe reinforcement learning by offline demonstration
Safe Reinforcement Learning (RL) aims to find a policy that achieves high rewards while
satisfying cost constraints. When learning from scratch, safe RL agents tend to be overly …
satisfying cost constraints. When learning from scratch, safe RL agents tend to be overly …
Residual q-learning: Offline and online policy customization without value
Imitation Learning (IL) is a widely used framework for learning imitative behavior from
demonstrations. It is especially appealing for solving complex real-world tasks where …
demonstrations. It is especially appealing for solving complex real-world tasks where …
Adaptive prediction ensemble: Improving out-of-distribution generalization of motion forecasting
Deep learning-based trajectory prediction models for autonomous driving often struggle with
generalization to out-of-distribution (OOD) scenarios, sometimes performing worse than …
generalization to out-of-distribution (OOD) scenarios, sometimes performing worse than …
Odice: Revealing the mystery of distribution correction estimation via orthogonal-gradient update
In this study, we investigate the DIstribution Correction Estimation (DICE) methods, an
important line of work in offline reinforcement learning (RL) and imitation learning (IL). DICE …
important line of work in offline reinforcement learning (RL) and imitation learning (IL). DICE …
Domain: Mildly conservative model-based offline reinforcement learning
Model-based reinforcement learning (RL), which learns environment model from offline
dataset and generates more out-of-distribution model data, has become an effective …
dataset and generates more out-of-distribution model data, has become an effective …