V2X-Sim: Multi-agent collaborative perception dataset and benchmark for autonomous driving

Y Li, D Ma, Z An, Z Wang, Y Zhong… - IEEE Robotics and …, 2022 - ieeexplore.ieee.org
Vehicle-to-everything (V2X) communication techniques enable the collaboration between
vehicles and many other entities in the neighboring environment, which could fundamentally …

Drive like a human: Rethinking autonomous driving with large language models

D Fu, X Li, L Wen, M Dou, P Cai… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this paper, we explore the potential of using a large language model (LLM) to understand
the driving environment in a human-like manner and analyze its ability to reason, interpret …

Spatio-temporal image representation and deep-learning-based decision framework for automated vehicles

S Cheng, B Yang, Z Wang… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Driving maneuver decision-making is critical to the development and mass deployment of
automated vehicles (AVs). The prevailing approaches are stuck with either specific …

Lane-attention: Predicting vehicles' moving trajectories by learning their attention over lanes

J Pan, H Sun, K Xu, Y Jiang, X Xiao… - 2020 IEEE/RSJ …, 2020 - ieeexplore.ieee.org
Accurately forecasting the future movements of surrounding vehicles is essential for safe
and efficient operations of autonomous driving cars. This task is difficult because a vehicle's …

St-p3: End-to-end vision-based autonomous driving via spatial-temporal feature learning

S Hu, L Chen, P Wu, H Li, J Yan, D Tao - European Conference on …, 2022 - Springer
Many existing autonomous driving paradigms involve a multi-stage discrete pipeline of
tasks. To better predict the control signals and enhance user safety, an end-to-end approach …

Towards compact autonomous driving perception with balanced learning and multi-sensor fusion

O Natan, J Miura - IEEE Transactions on Intelligent …, 2022 - ieeexplore.ieee.org
We present a novel compact deep multi-task learning model to handle various autonomous
driving perception tasks in one forward pass. The model performs multiple views of semantic …

Licanet: Further enhancement of joint perception and motion prediction based on multi-modal fusion

YH Khalil, HT Mouftah - IEEE Open Journal of Intelligent …, 2022 - ieeexplore.ieee.org
The safety and reliability of autonomous driving pivots on the accuracy of perception and
motion prediction pipelines, which reckons primarily on the sensors deployed onboard …

Exploring imitation learning for autonomous driving with feedback synthesizer and differentiable rasterization

J Zhou, R Wang, X Liu, Y Jiang, S Jiang… - 2021 IEEE/RSJ …, 2021 - ieeexplore.ieee.org
We present a learning-based planner that aims to robustly drive a vehicle by mimicking
human drivers' driving behavior. We leverage a mid-to-mid approach that allows us to …

Multi-modal fusion transformer for end-to-end autonomous driving

A Prakash, K Chitta, A Geiger - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
How should representations from complementary sensors be integrated for autonomous
driving? Geometry-based sensor fusion has shown great promise for perception tasks such …

Spatiotemporal scene-graph embedding for autonomous vehicle collision prediction

AV Malawade, SY Yu, B Hsu… - IEEE Internet of …, 2022 - ieeexplore.ieee.org
In autonomous vehicles (AVs), early warning systems rely on collision prediction to ensure
occupant safety. However, state-of-the-art methods using deep convolutional networks …