PanopticNDT: Efficient and Robust Panoptic Mapping

As the application scenarios of mobile robots are getting more complex and challenging, scene understanding becomes increasingly crucial. A mobile robot that is supposed to operate autonomously in indoor environments must have precise knowledge about what objects are present, where they are, what their spatial extent is, and how they can be reached; i.e., information about free space is also crucial. Panoptic mapping is a powerful instrument providing such information. However, building 3D panoptic maps with high spatial resolution is challenging on mobile robots, given their limited computing capabilities. In this paper, we propose PanopticNDT - an efficient and robust panoptic mapping approach based on occupancy normal distribution transform (NDT) mapping. We evaluate our approach on the publicly available datasets Hypersim and ScanNetV2. The results reveal that our approach can represent panoptic information at a higher level of detail than other state-of-the-art approaches while enabling real-time panoptic mapping on mobile robots. Finally, we prove the real-world applicability of PanopticNDT with qualitative results in a domestic application.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Panoptic Segmentation Hypersim EMSANet (2x ResNet-34 NBt1D) PQ 34.95 # 1
PQ (test) 29.77 # 1
mIoU 49.12 # 1
mIoU (test) 44.66 # 1
Semantic Segmentation Hypersim EMSANet (2x ResNet-34 NBt1D) mIoU 49.74 # 1
mIoU (test) 46.66 # 1
3D Semantic Segmentation Hypersim SemanticNDT (10cm) mIoU 44.31 # 2
mIoU (test) 44.8 # 2
3D Panoptic Segmentation Hypersim PanopticNDT (10cm) PQ 31.08 # 1
PQ (test) 27.54 # 1
mIoU 44.56 # 1
mIoU (test) 45.2 # 1
3D Semantic Segmentation Hypersim PanopticNDT (10cm) mIoU 45.43 # 1
mIoU (test) 45.34 # 1
Panoptic Segmentation NYU Depth v2 EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) PQ 51.15 # 1
Scene Classification (unified classes) NYU Depth v2 EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) Balanced Accuracy 73.33 # 2
Semantic Segmentation NYU Depth v2 EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) Mean IoU 59.02 # 8
Semantic Segmentation ScanNet PanopticNDT (10cm) test mIoU 68.1 # 23
val mIoU 68.39 # 30
Semantic Segmentation ScanNetV2 EMSANet (2x ResNet-34 NBt1D, PanopticNDT version) Mean IoU 60.0% # 2
Mean IoU (val) 70.99% # 1
Mean IoU (test) 60.0% # 1
2D Panoptic Segmentation ScanNetV2 EMSANet (2x ResNet-34 NBt1D, PanopticNDT version) PQ 58.22 # 1
Panoptic Segmentation ScanNetV2 PanopticNDT (10cm) PQ 59.19 # 2
Panoptic Segmentation (PanopticNDT instances) SUN-RGBD EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) PQ 52.48 # 1
Scene Classification (unified classes) SUN-RGBD EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) Balanced Accuracy 58.55 # 1
Semantic Segmentation SUN-RGBD EMSANet (2x ResNet-34 NBt1D, PanopticNDT version, finetuned) Mean IoU 50.86% # 15

Methods


No methods listed for this paper. Add relevant methods here