Arcap: Collecting high-quality human demonstrations for robot learning with augmented reality feedback

S Chen, C Wang, K Nguyen, L Fei-Fei… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent progress in imitation learning from human demonstrations has shown promising
results in teaching robots manipulation skills. To further scale up training datasets, recent …

Generalizable humanoid manipulation with improved 3d diffusion policies

Y Ze, Z Chen, W Wang, T Chen, X He, Y Yuan… - arxiv preprint arxiv …, 2024 - arxiv.org
Humanoid robots capable of autonomous operation in diverse environments have long
been a goal for roboticists. However, autonomous manipulation by humanoid robots has …

Vividex: Learning vision-based dexterous manipulation from human videos

Z Chen, S Chen, E Arlaud, I Laptev… - arxiv preprint arxiv …, 2024 - arxiv.org
In this work, we aim to learn a unified vision-based policy for multi-fingered robot hands to
manipulate a variety of objects in diverse poses. Though prior work has shown benefits of …

Egomimic: Scaling imitation learning via egocentric video

S Kareer, D Patel, R Punamiya, P Mathur… - arxiv preprint arxiv …, 2024 - arxiv.org
The scale and diversity of demonstration data required for imitation learning is a significant
challenge. We present EgoMimic, a full-stack framework which scales manipulation via …

Active vision might be all you need: Exploring active vision in bimanual robotic manipulation

I Chuang, A Lee, D Gao, I Soltani - arxiv preprint arxiv:2409.17435, 2024 - arxiv.org
Imitation learning has demonstrated significant potential in performing high-precision
manipulation tasks using visual feedback from cameras. However, it is common practice in …

ForceMimic: Force-Centric Imitation Learning with Force-Motion Capture System for Contact-Rich Manipulation

W Liu, J Wang, Y Wang, W Wang, C Lu - arxiv preprint arxiv:2410.07554, 2024 - arxiv.org
In most contact-rich manipulation tasks, humans apply time-varying forces to the target
object, compensating for inaccuracies in the vision-guided hand trajectory. However, current …

Canonical representation and force-based pretraining of 3d tactile for dexterous visuo-tactile policy learning

T Wu, J Li, J Zhang, M Wu, H Dong - arxiv preprint arxiv:2409.17549, 2024 - arxiv.org
Tactile sensing plays a vital role in enabling robots to perform fine-grained, contact-rich
tasks. However, the high dimensionality of tactile data, due to the large coverage on …

Bridging the Human to Robot Dexterity Gap through Object-Oriented Rewards

I Guzey, Y Dai, G Savva, R Bhirangi, L Pinto - arxiv preprint arxiv …, 2024 - arxiv.org
Training robots directly from human videos is an emerging area in robotics and computer
vision. While there has been notable progress with two-fingered grippers, learning …

DexForce: Extracting Force-informed Actions from Kinesthetic Demonstrations for Dexterous Manipulation

C Chen, Z Yu, H Choi, M Cutkosky, J Bohg - arxiv preprint arxiv …, 2025 - arxiv.org
Imitation learning requires high-quality demonstrations consisting of sequences of state-
action pairs. For contact-rich dexterous manipulation tasks that require fine-grained …

Learning to Look Around: Enhancing Teleoperation and Learning with a Human-like Actuated Neck

B Sen, M Wang, N Thakur, A Agarwal… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce a teleoperation system that integrates a 5 DOF actuated neck, designed to
replicate natural human head movements and perception. By enabling behaviors like …