PETR: Position Embedding Transformation for Multi-View 3D Object Detection

10 Mar 2022  ·  Yingfei Liu, Tiancai Wang, Xiangyu Zhang, Jian Sun ·

In this paper, we develop position embedding transformation (PETR) for multi-view 3D object detection. PETR encodes the position information of 3D coordinates into image features, producing the 3D position-aware features. Object query can perceive the 3D position-aware features and perform end-to-end object detection. PETR achieves state-of-the-art performance (50.4% NDS and 44.1% mAP) on standard nuScenes dataset and ranks 1st place on the benchmark. It can serve as a simple yet strong baseline for future research. Code is available at \url{https://github.com/megvii-research/PETR}.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
3D Object Detection 3D Object Detection on Argoverse2 Camera Only PETR Average mAP 17.6 # 3
Robust Camera Only 3D Object Detection nuScenes-C PETR (vov) mean Corruption Error (mCE) 100.69 # 12
mean Resilience Rate (mRR) 65.03 # 8
Robust Camera Only 3D Object Detection nuScenes-C PETR (r50) mean Corruption Error (mCE) 111.01 # 19
mean Resilience Rate (mRR) 61.26 # 12

Results from Other Papers


Task Dataset Model Metric Name Metric Value Rank Source Paper Compare
3D Object Detection TruckScenes PETR NDS 12.1 # 3
mAP 2.2 # 4

Methods


No methods listed for this paper. Add relevant methods here