Manigaussian: Dynamic gaussian splatting for multi-task robotic manipulation

G Lu, S Zhang, Z Wang, C Liu, J Lu, Y Tang - European Conference on …, 2024 - Springer
Performing language-conditioned robotic manipulation tasks in unstructured environments
is highly demanded for general intelligent robots. Conventional robotic manipulation …

Bunny-visionpro: Real-time bimanual dexterous teleoperation for imitation learning

R Ding, Y Qin, J Zhu, C Jia, S Yang, R Yang… - arxiv preprint arxiv …, 2024 - arxiv.org
Teleoperation is a crucial tool for collecting human demonstrations, but controlling robots
with bimanual dexterous hands remains a challenge. Existing teleoperation systems …

Affordancellm: Grounding affordance from vision language models

S Qian, W Chen, M Bai, X Zhou… - Proceedings of the …, 2024 - openaccess.thecvf.com
Affordance grounding refers to the task of finding the area of an object with which one can
interact. It is a fundamental but challenging task as a successful solution requires the …

Manipulate-anything: Automating real-world robots using vision-language models

J Duan, W Yuan, W Pumacay, YR Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large-scale endeavors like and widespread community efforts such as Open-X-Embodiment
have contributed to growing the scale of robot demonstration data. However, there is still an …

Hierarchical Diffusion Policy for Kinematics-Aware Multi-Task Robotic Manipulation

X Ma, S Patidar, I Haughton… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract This paper introduces Hierarchical Diffusion Policy (HDP) a hierarchical agent for
multi-task robotic manipulation. HDP factorises a manipulation policy into a hierarchical …

READ: Retrieval-Enhanced Asymmetric Diffusion for Motion Planning

T Oba, M Walter, N Ukita - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
Abstract This paper proposes Retrieval-Enhanced Asymmetric Diffusion (READ) for image-
based robot motion planning. Given an image of the scene READ retrieves an initial motion …

Deep generative models in robotics: A survey on learning from multimodal demonstrations

J Urain, A Mandlekar, Y Du, M Shafiullah, D Xu… - arxiv preprint arxiv …, 2024 - arxiv.org
Learning from Demonstrations, the field that proposes to learn robot behavior models from
data, is gaining popularity with the emergence of deep generative models. Although the …

Learning to manipulate anywhere: A visual generalizable framework for reinforcement learning

Z Yuan, T Wei, S Cheng, G Zhang, Y Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Can we endow visuomotor robots with generalization capabilities to operate in diverse open-
world scenarios? In this paper, we propose\textbf {Maniwhere}, a generalizable framework …

3d diffuser actor: Policy diffusion with 3d scene representations

TW Ke, N Gkanatsios, K Fragkiadaki - arxiv preprint arxiv:2402.10885, 2024 - arxiv.org
We marry diffusion policies and 3D scene representations for robot manipulation. Diffusion
policies learn the action distribution conditioned on the robot and environment state using …

Fourier transporter: Bi-equivariant robotic manipulation in 3d

H Huang, O Howell, D Wang, X Zhu, R Walters… - arxiv preprint arxiv …, 2024 - arxiv.org
Many complex robotic manipulation tasks can be decomposed as a sequence of pick and
place actions. Training a robotic agent to learn this sequence over many different starting …