Uni-to-Multi Modal Knowledge Distillation for Bidirectional LiDAR-Camera Semantic Segmentation

T Sun, Z Zhang, X Tan, Y Peng, Y Qu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Combining LiDAR points and images for robust semantic segmentation has shown great
potential. However, the heterogeneity between the two modalities (eg the density, the field of …

LPFormer: LiDAR pose estimation transformer with multi-task network

D Ye, Y **e, W Chen, Z Zhou, L Ge… - … on Robotics and …, 2024 - ieeexplore.ieee.org
Due to the difficulty of acquiring large-scale 3D human keypoint annotation, previous
methods for 3D human pose estimation (HPE) have often relied on 2D image features and …

Panoptic Perception for Autonomous Driving: A Survey

Y Li, L Xu - arxiv preprint arxiv:2408.15388, 2024 - arxiv.org
Panoptic perception represents a forefront advancement in autonomous driving technology,
unifying multiple perception tasks into a singular, cohesive framework to facilitate a thorough …

Small, versatile and mighty: A range-view perception framework

Q Meng, X Wang, JB Wang, L Yan, K Wang - arxiv preprint arxiv …, 2024 - arxiv.org
Despite its compactness and information integrity, the range view representation of LiDAR
data rarely occurs as the first choice for 3D perception tasks. In this work, we further push the …

Cocoon: Robust Multi-Modal Perception with Uncertainty-Aware Sensor Fusion

M Cho, Y Cao, J Sun, Q Zhang, M Pavone… - arxiv preprint arxiv …, 2024 - arxiv.org
An important paradigm in 3D object detection is the use of multiple modalities to enhance
accuracy in both normal and challenging conditions, particularly for long-tail scenarios. To …

A Point-Based Approach to Efficient LiDAR Multi-Task Perception

C Lang, A Braun, L Schillingmann, A Valada - arxiv preprint arxiv …, 2024 - arxiv.org
Multi-task networks can potentially improve performance and computational efficiency
compared to single-task networks, facilitating online deployment. However, current multi …

3D-JEPA: A Joint Embedding Predictive Architecture for 3D Self-Supervised Representation Learning

N Hu, H Cheng, Y **e, S Li, J Zhu - arxiv preprint arxiv:2409.15803, 2024 - arxiv.org
Invariance-based and generative methods have shown a conspicuous performance for 3D
self-supervised representation learning (SSRL). However, the former relies on hand-crafted …

LiSD: An Efficient Multi-Task Learning Framework for LiDAR Segmentation and Detection

J Xu, S Zuo, C Wei, W Zhou - arxiv preprint arxiv:2406.07023, 2024 - arxiv.org
With the rapid proliferation of autonomous driving, there has been a heightened focus on the
research of lidar-based 3D semantic segmentation and object detection methodologies …