Delving into the devils of bird's-eye-view perception: A review, evaluation and recipe

H Li, C Sima, J Dai, W Wang, L Lu… - … on Pattern Analysis …, 2023 - ieeexplore.ieee.org
Learning powerful representations in bird's-eye-view (BEV) for perception tasks is trending
and drawing extensive attention both from industry and academia. Conventional …

3D object detection for autonomous driving: A comprehensive survey

J Mao, S Shi, X Wang, H Li - International Journal of Computer Vision, 2023 - Springer
Autonomous driving, in recent years, has been receiving increasing attention for its potential
to relieve drivers' burdens and improve the safety of driving. In modern autonomous driving …

Surroundocc: Multi-camera 3d occupancy prediction for autonomous driving

Y Wei, L Zhao, W Zheng, Z Zhu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract 3D scene understanding plays a vital role in vision-based autonomous driving.
While most existing methods focus on 3D object detection, they have difficulty describing …

Bevdepth: Acquisition of reliable depth for multi-view 3d object detection

Y Li, Z Ge, G Yu, J Yang, Z Wang, Y Shi… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
In this research, we propose a new 3D object detector with a trustworthy depth estimation,
dubbed BEVDepth, for camera-based Bird's-Eye-View~(BEV) 3D object detection. Our work …

Bevfusion: Multi-task multi-sensor fusion with unified bird's-eye view representation

Z Liu, H Tang, A Amini, X Yang, H Mao… - … on robotics and …, 2023 - ieeexplore.ieee.org
Multi-sensor fusion is essential for an accurate and reliable autonomous driving system.
Recent approaches are based on point-level fusion: augmenting the LiDAR point cloud with …

Bevformer: Learning bird's-eye-view representation from multi-camera images via spatiotemporal transformers

Z Li, W Wang, H Li, E Xie, C Sima, T Lu, Y Qiao… - European conference on …, 2022 - Springer
Abstract 3D visual perception tasks, including 3D detection and map segmentation based on
multi-camera images, are essential for autonomous driving systems. In this work, we present …

Petr: Position embedding transformation for multi-view 3d object detection

Y Liu, T Wang, X Zhang, J Sun - European Conference on Computer …, 2022 - Springer
In this paper, we develop position embedding transformation (PETR) for multi-view 3D
object detection. PETR encodes the position information of 3D coordinates into image …

Petrv2: A unified framework for 3d perception from multi-camera images

Y Liu, J Yan, F Jia, S Li, A Gao… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this paper, we propose PETRv2, a unified framework for 3D perception from multi-view
images. Based on PETR, PETRv2 explores the effectiveness of temporal modeling, which …

Bevfusion: A simple and robust lidar-camera fusion framework

T Liang, H Xie, K Yu, Z Xia, Z Lin… - Advances in …, 2022 - proceedings.neurips.cc
Fusing the camera and LiDAR information has become a de-facto standard for 3D object
detection tasks. Current methods rely on point clouds from the LiDAR sensor as queries to …

Bevdet: High-performance multi-camera 3d object detection in bird-eye-view

J Huang, G Huang, Z Zhu, Y Ye, D Du - arXiv preprint arXiv:2112.11790, 2021 - arxiv.org
Autonomous driving perceives its surroundings for decision making, which is one of the most
complex scenarios in visual perception. The success of paradigm innovation in solving the …