PETR: Position Embedding Transformation for Multi-View 3D Object Detection

10 Mar 2022  ·  Yingfei Liu, Tiancai Wang, Xiangyu Zhang, Jian Sun ·

In this paper, we develop position embedding transformation (PETR) for multi-view 3D object detection. PETR encodes the position information of 3D coordinates into image features, producing the 3D position-aware features. Object query can perceive the 3D position-aware features and perform end-to-end object detection. PETR achieves state-of-the-art performance (50.4% NDS and 44.1% mAP) on standard nuScenes dataset and ranks 1st place on the benchmark. It can serve as a simple yet strong baseline for future research. Code is available at \url{https://github.com/megvii-research/PETR}.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
3D Object Detection 3D Object Detection on Argoverse2 Camera Only PETR Average mAP 17.6 # 3
Robust Camera Only 3D Object Detection nuScenes-C PETR (vov) mean Corruption Error (mCE) 100.69 # 12
mean Resilience Rate (mRR) 65.03 # 8
Robust Camera Only 3D Object Detection nuScenes-C PETR (r50) mean Corruption Error (mCE) 111.01 # 19
mean Resilience Rate (mRR) 61.26 # 12

Methods


No methods listed for this paper. Add relevant methods here