Multi-modal 3d object detection in autonomous driving: A survey and taxonomy

L Wang, X Zhang, Z Song, J Bi, G Zhang… - IEEE Transactions …, 2023 - ieeexplore.ieee.org
Autonomous vehicles require constant environmental perception to obtain the distribution of
obstacles to achieve safe driving. Specifically, 3D object detection is a vital functional …

Delving into the devils of bird's-eye-view perception: A review, evaluation and recipe

H Li, C Sima, J Dai, W Wang, L Lu… - … on Pattern Analysis …, 2023 - ieeexplore.ieee.org
Learning powerful representations in bird's-eye-view (BEV) for perception tasks is trending
and drawing extensive attention both from industry and academia. Conventional …

Occ3d: A large-scale 3d occupancy prediction benchmark for autonomous driving

X Tian, T Jiang, L Yun, Y Mao, H Yang… - Advances in …, 2024 - proceedings.neurips.cc
Robotic perception requires the modeling of both 3D geometry and semantics. Existing
methods typically focus on estimating 3D bounding boxes, neglecting finer geometric details …

Bevfusion: Multi-task multi-sensor fusion with unified bird's-eye view representation

Z Liu, H Tang, A Amini, X Yang, H Mao… - … on robotics and …, 2023 - ieeexplore.ieee.org
Multi-sensor fusion is essential for an accurate and reliable autonomous driving system.
Recent approaches are based on point-level fusion: augmenting the LiDAR point cloud with …

Bevfusion: A simple and robust lidar-camera fusion framework

T Liang, H Xie, K Yu, Z Xia, Z Lin… - Advances in …, 2022 - proceedings.neurips.cc
Fusing the camera and LiDAR information has become a de-facto standard for 3D object
detection tasks. Current methods rely on point clouds from the LiDAR sensor as queries to …

Drivinggaussian: Composite gaussian splatting for surrounding dynamic autonomous driving scenes

X Zhou, Z Lin, X Shan, Y Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
We present DrivingGaussian an efficient and effective framework for surrounding dynamic
autonomous driving scenes. For complex scenes with moving objects we first sequentially …

Virtual sparse convolution for multimodal 3d object detection

H Wu, C Wen, S Shi, X Li… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract Recently, virtual/pseudo-point-based 3D object detection that seamlessly fuses
RGB images and LiDAR data by depth completion has gained great attention. However …

Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection

Y Li, AW Yu, T Meng, B Caine… - Proceedings of the …, 2022 - openaccess.thecvf.com
Lidars and cameras are critical sensors that provide complementary information for 3D
detection in autonomous driving. While prevalent multi-modal methods simply decorate raw …

Unifying voxel-based representation with transformer for 3d object detection

Y Li, Y Chen, X Qi, Z Li, J Sun… - Advances in Neural …, 2022 - proceedings.neurips.cc
In this work, we present a unified framework for multi-modality 3D object detection, named
UVTR. The proposed method aims to unify multi-modality representations in the voxel space …

Transfuser: Imitation with transformer-based sensor fusion for autonomous driving

K Chitta, A Prakash, B Jaeger, Z Yu… - … on Pattern Analysis …, 2022 - ieeexplore.ieee.org
How should we integrate representations from complementary sensors for autonomous
driving? Geometry-based fusion has shown promise for perception (eg, object detection …