Cylinder3D: An Effective 3D Framework for Driving-scene LiDAR Semantic Segmentation

4 Aug 2020  ·  Hui Zhou, Xinge Zhu, Xiao Song, Yuexin Ma, Zhe Wang, Hongsheng Li, Dahua Lin ·

State-of-the-art methods for large-scale driving-scene LiDAR semantic segmentation often project and process the point clouds in the 2D space. The projection methods includes spherical projection, bird-eye view projection, etc. Although this process makes the point cloud suitable for the 2D CNN-based networks, it inevitably alters and abandons the 3D topology and geometric relations. A straightforward solution to tackle the issue of 3D-to-2D projection is to keep the 3D representation and process the points in the 3D space. In this work, we first perform an in-depth analysis for different representations and backbones in 2D and 3D spaces, and reveal the effectiveness of 3D representations and networks on LiDAR segmentation. Then, we develop a 3D cylinder partition and a 3D cylinder convolution based framework, termed as Cylinder3D, which exploits the 3D topology relations and structures of driving-scene point clouds. Moreover, a dimension-decomposition based context modeling module is introduced to explore the high-rank context information in point clouds in a progressive manner. We evaluate the proposed model on a large-scale driving-scene dataset, i.e. SematicKITTI. Our method achieves state-of-the-art performance and outperforms existing methods by 6% in terms of mIoU.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
LIDAR Semantic Segmentation nuScenes Cylinder3D++ test mIoU 0.78 # 11
3D Object Detection nuScenes Reconfig PP v3 NDS 0.59 # 179
mAP 0.49 # 195
mATE 0.33 # 202
mASE 0.24 # 213
mAOE 0.44 # 88
mAVE 0.27 # 238
mAAE 0.24 # 38

Methods