3D object detection for autonomous driving: A comprehensive survey

J Mao, S Shi, X Wang, H Li - International Journal of Computer Vision, 2023‏ - Springer
Autonomous driving, in recent years, has been receiving increasing attention for its potential
to relieve drivers' burdens and improve the safety of driving. In modern autonomous driving …

Deep learning for 3d point clouds: A survey

Y Guo, H Wang, Q Hu, H Liu, L Liu… - IEEE transactions on …, 2020‏ - ieeexplore.ieee.org
Point cloud learning has lately attracted increasing attention due to its wide applications in
many areas, such as computer vision, autonomous driving, and robotics. As a dominating …

Planning-oriented autonomous driving

Y Hu, J Yang, L Chen, K Li, C Sima… - Proceedings of the …, 2023‏ - openaccess.thecvf.com
Modern autonomous driving system is characterized as modular tasks in sequential order,
ie, perception, prediction, and planning. In order to perform a wide diversity of tasks and …

Bevformer: learning bird's-eye-view representation from lidar-camera via spatiotemporal transformers

Z Li, W Wang, H Li, E **e, C Sima, T Lu… - IEEE Transactions on …, 2024‏ - ieeexplore.ieee.org
Multi-modality fusion strategy is currently the de-facto most competitive solution for 3D
perception tasks. In this work, we present a new framework termed BEVFormer, which learns …

Bytetrack: Multi-object tracking by associating every detection box

Y Zhang, P Sun, Y Jiang, D Yu, F Weng, Z Yuan… - European conference on …, 2022‏ - Springer
Multi-object tracking (MOT) aims at estimating bounding boxes and identities of objects in
videos. Most methods obtain identities by associating detection boxes whose scores are …

Transfuser: Imitation with transformer-based sensor fusion for autonomous driving

K Chitta, A Prakash, B Jaeger, Z Yu… - … on Pattern Analysis …, 2022‏ - ieeexplore.ieee.org
How should we integrate representations from complementary sensors for autonomous
driving? Geometry-based fusion has shown promise for perception (eg, object detection …

Multi-modal fusion transformer for end-to-end autonomous driving

A Prakash, K Chitta, A Geiger - Proceedings of the IEEE …, 2021‏ - openaccess.thecvf.com
How should representations from complementary sensors be integrated for autonomous
driving? Geometry-based sensor fusion has shown great promise for perception tasks such …

St-p3: End-to-end vision-based autonomous driving via spatial-temporal feature learning

S Hu, L Chen, P Wu, H Li, J Yan, D Tao - European Conference on …, 2022‏ - Springer
Many existing autonomous driving paradigms involve a multi-stage discrete pipeline of
tasks. To better predict the control signals and enhance user safety, an end-to-end approach …

Beverse: Unified perception and prediction in birds-eye-view for vision-centric autonomous driving

Y Zhang, Z Zhu, W Zheng, J Huang, G Huang… - arxiv preprint arxiv …, 2022‏ - arxiv.org
In this paper, we present BEVerse, a unified framework for 3D perception and prediction
based on multi-camera systems. Unlike existing studies focusing on the improvement of …

Learning from all vehicles

D Chen, P Krähenbühl - … of the IEEE/CVF Conference on …, 2022‏ - openaccess.thecvf.com
In this paper, we present a system to train driving policies from experiences collected not just
from the ego-vehicle, but all vehicles that it observes. This system uses the behaviors of …