Towards Deep Radar Perception for Autonomous Driving: Datasets, Methods, and Challenges

Yi Zhou, Lulu Liu, Haocheng Zhao, Miguel López-Benítez, Limin Yu, Yutao Yue*

*Corresponding author for this work

Research output: Contribution to journalReview articlepeer-review

48 Citations (Scopus)

Abstract

With recent developments, the performance of automotive radar has improved significantly. The next generation of 4D radar can achieve imaging capability in the form of high-resolution point clouds. In this context, we believe that the era of deep learning for radar perception has arrived. However, studies on radar deep learning are spread across different tasks, and a holistic overview is lacking. This review paper attempts to provide a big picture of the deep radar perception stack, including signal processing, datasets, labelling, data augmentation, and downstream tasks such as depth and velocity estimation, object detection, and sensor fusion. For these tasks, we focus on explaining how the network structure is adapted to radar domain knowledge. In particular, we summarise three overlooked challenges in deep radar perception, including multi-path effects, uncertainty problems, and adverse weather effects, and present some attempts to solve them.

Original languageEnglish
Article number4208
JournalSensors
Volume22
Issue number11
DOIs
Publication statusPublished - 1 Jun 2022

Keywords

  • automotive radars
  • autonomous driving
  • deep learning
  • multi-sensor fusion
  • object detection
  • radar signal processing

Fingerprint

Dive into the research topics of 'Towards Deep Radar Perception for Autonomous Driving: Datasets, Methods, and Challenges'. Together they form a unique fingerprint.

Cite this