3D-MiniNet: Learning a 2D Representation from Point Clouds for Fast and Efficient 3D LIDAR Semantic Segmentation

25 Feb 2020  ·  Iñigo Alonso, Luis Riazuelo, Luis Montesano, Ana C. Murillo ·

LIDAR semantic segmentation, which assigns a semantic label to each 3D point measured by the LIDAR, is becoming an essential task for many robotic applications such as autonomous driving. Fast and efficient semantic segmentation methods are needed to match the strong computational and temporal restrictions of many of these real-world applications. This work presents 3D-MiniNet, a novel approach for LIDAR semantic segmentation that combines 3D and 2D learning layers. It first learns a 2D representation from the raw points through a novel projection which extracts local and global information from the 3D data. This representation is fed to an efficient 2D Fully Convolutional Neural Network (FCNN) that produces a 2D semantic segmentation. These 2D semantic labels are re-projected back to the 3D space and enhanced through a post-processing module. The main novelty in our strategy relies on the projection learning module. Our detailed ablation study shows how each component contributes to the final performance of 3D-MiniNet. We validate our approach on well known public benchmarks (SemanticKITTI and KITTI), where 3D-MiniNet gets state-of-the-art results while being faster and more parameter-efficient than previous methods.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Real-Time 3D Semantic Segmentation SemanticKITTI 3D-MiniNet-tiny Speed (FPS) 98 # 1
mIoU 46.9 # 3
Parameters (M) 0.44 # 1
Real-Time 3D Semantic Segmentation SemanticKITTI 3D-MiniNet Speed (FPS) 28 # 3
mIoU 55.8 # 1
Parameters (M) 3.97 # 3
3D Semantic Segmentation SemanticKITTI 3D-MiniNet mIoU 55.8% # 16


No methods listed for this paper. Add relevant methods here