Aligning cyber space with physical world: A comprehensive survey on embodied ai

Y Liu, W Chen, Y Bai, X Liang, G Li, W Gao… - arxiv preprint arxiv …, 2024 - arxiv.org
Embodied Artificial Intelligence (Embodied AI) is crucial for achieving Artificial General
Intelligence (AGI) and serves as a foundation for various applications that bridge cyberspace …

Binding touch to everything: Learning unified multimodal tactile representations

F Yang, C Feng, Z Chen, H Park… - Proceedings of the …, 2024 - openaccess.thecvf.com
The ability to associate touch with other modalities has huge implications for humans and
computational systems. However multimodal learning with touch remains challenging due to …

Tactile-augmented radiance fields

Y Dou, F Yang, Y Liu, A Loquercio… - Proceedings of the …, 2024 - openaccess.thecvf.com
We present a scene representation that brings vision and touch into a shared 3D space
which we call a tactile-augmented radiance field. This representation capitalizes on two key …

Generating visual scenes from touch

F Yang, J Zhang, A Owens - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
An emerging line of work has sought to generate plausible imagery from touch. Existing
approaches, however, tackle only narrow aspects of the visuo-tactile synthesis problem, and …

Multiply: A multisensory object-centric embodied large language model in 3d world

Y Hong, Z Zheng, P Chen, Y Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Human beings possess the capability to multiply a melange of multisensory cues while
actively exploring and interacting with the 3D world. Current multi-modal large language …

Maniwav: Learning robot manipulation from in-the-wild audio-visual data

Z Liu, C Chi, E Cousineau, N Kuppuswamy… - … Conference on Robot …, 2024 - openreview.net
Audio signals provide rich information for the robot interaction and object properties through
contact. These information can surprisingly ease the learning of contact-rich robot …

Octopi: Object property reasoning with large tactile-language models

S Yu, K Lin, A **ao, J Duan, H Soh - arxiv preprint arxiv:2405.02794, 2024 - arxiv.org
Physical reasoning is important for effective robot manipulation. Recent work has
investigated both vision and language modalities for physical reasoning; vision can reveal …

Snap-it, tap-it, splat-it: Tactile-informed 3d gaussian splatting for reconstructing challenging surfaces

M Comi, A Tonioni, M Yang, J Tremblay… - arxiv preprint arxiv …, 2024 - arxiv.org
Touch and vision go hand in hand, mutually enhancing our ability to understand the world.
From a research perspective, the problem of mixing touch and vision is underexplored and …

Robopack: Learning tactile-informed dynamics models for dense packing

B Ai, S Tian, H Shi, Y Wang, C Tan, Y Li… - arxiv preprint arxiv …, 2024 - arxiv.org
Tactile feedback is critical for understanding the dynamics of both rigid and deformable
objects in many manipulation tasks, such as non-prehensile manipulation and dense …

Touch2touch: Cross-modal tactile generation for object manipulation

S Rodriguez, Y Dou, M Oller, A Owens… - arxiv preprint arxiv …, 2024 - arxiv.org
Today's touch sensors come in many shapes and sizes. This has made it challenging to
develop general-purpose touch processing methods since models are generally tied to one …