Multi-modal 3d object detection in autonomous driving: A survey and taxonomy
Autonomous vehicles require constant environmental perception to obtain the distribution of
obstacles to achieve safe driving. Specifically, 3D object detection is a vital functional …
obstacles to achieve safe driving. Specifically, 3D object detection is a vital functional …
Delving into the devils of bird's-eye-view perception: A review, evaluation and recipe
Learning powerful representations in bird's-eye-view (BEV) for perception tasks is trending
and drawing extensive attention both from industry and academia. Conventional …
and drawing extensive attention both from industry and academia. Conventional …
Bevfusion: Multi-task multi-sensor fusion with unified bird's-eye view representation
Multi-sensor fusion is essential for an accurate and reliable autonomous driving system.
Recent approaches are based on point-level fusion: augmenting the LiDAR point cloud with …
Recent approaches are based on point-level fusion: augmenting the LiDAR point cloud with …
Transfusion: Robust lidar-camera fusion for 3d object detection with transformers
LiDAR and camera are two important sensors for 3D object detection in autonomous driving.
Despite the increasing popularity of sensor fusion in this field, the robustness against inferior …
Despite the increasing popularity of sensor fusion in this field, the robustness against inferior …
Bevfusion: A simple and robust lidar-camera fusion framework
Fusing the camera and LiDAR information has become a de-facto standard for 3D object
detection tasks. Current methods rely on point clouds from the LiDAR sensor as queries to …
detection tasks. Current methods rely on point clouds from the LiDAR sensor as queries to …
Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection
Lidars and cameras are critical sensors that provide complementary information for 3D
detection in autonomous driving. While prevalent multi-modal methods simply decorate raw …
detection in autonomous driving. While prevalent multi-modal methods simply decorate raw …
Unifying voxel-based representation with transformer for 3d object detection
In this work, we present a unified framework for multi-modality 3D object detection, named
UVTR. The proposed method aims to unify multi-modality representations in the voxel space …
UVTR. The proposed method aims to unify multi-modality representations in the voxel space …
V2x-vit: Vehicle-to-everything cooperative perception with vision transformer
In this paper, we investigate the application of Vehicle-to-Everything (V2X) communication to
improve the perception performance of autonomous vehicles. We present a robust …
improve the perception performance of autonomous vehicles. We present a robust …
Focal sparse convolutional networks for 3d object detection
Non-uniformed 3D sparse data, eg, point clouds or voxels in different spatial positions, make
contribution to the task of 3D object detection in different ways. Existing basic components in …
contribution to the task of 3D object detection in different ways. Existing basic components in …
Transfuser: Imitation with transformer-based sensor fusion for autonomous driving
How should we integrate representations from complementary sensors for autonomous
driving? Geometry-based fusion has shown promise for perception (eg, object detection …
driving? Geometry-based fusion has shown promise for perception (eg, object detection …