Exploiting Temporal Relations on Radar Perception for Autonomous Driving

CVPR 2022  ·  Peizhao Li, Pu Wang, Karl Berntorp, Hongfu Liu ·

We consider the object recognition problem in autonomous driving using automotive radar sensors. Comparing to Lidar sensors, radar is cost-effective and robust in all-weather conditions for perception in autonomous driving. However, radar signals suffer from low angular resolution and precision in recognizing surrounding objects. To enhance the capacity of automotive radar, in this work, we exploit the temporal information from successive ego-centric bird-eye-view radar image frames for radar object recognition. We leverage the consistency of an object's existence and attributes (size, orientation, etc.), and propose a temporal relational layer to explicitly model the relations between objects within successive radar images. In both object detection and multiple object tracking, we show the superiority of our method compared to several baseline approaches.

PDF Abstract CVPR 2022 PDF CVPR 2022 Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Multiple Object Tracking RADIATE TempoRadar MOTA 37.91 # 1
2D Object Detection RADIATE TempoRadar mAP@0.3 63.63±2.08 # 1

Methods


No methods listed for this paper. Add relevant methods here