Variational Context-Deformable ConvNets for Indoor Scene Parsing

Context information is critical for image semantic segmentation. Especially in indoor scenes, the large variation of object scales makes spatial-context an important factor for improving the segmentation performance. Thus, in this paper, we propose a novel variational context-deformable (VCD) module to learn adaptive receptive-field in a structured fashion. Different from standard ConvNets, which share fixed-size spatial context for all pixels, the VCD module learns a deformable spatial-context with the guidance of depth information: depth information provides clues for identifying real local neighborhoods. Specifically, adaptive Gaussian kernels are learned with the guidance of multimodal information. By multiplying the learned Gaussian kernel with standard convolution filters, the VCD module can aggregate flexible spatial context for each pixel during convolution. The main contributions of this work are as follows: 1) a novel VCD module is proposed, which exploits learnable Gaussian kernels to enable feature learning with structured adaptive-context; 2) variational Bayesian probabilistic modeling is introduced for the training of VCD module, which can make it continuous and more stable; 3) a perspective-aware guidance module is designed to take advantage of multi-modal information for RGB-D segmentation. We evaluate the proposed approach on three widely-used datasets, and the performance improvement has shown the effectiveness of the proposed method.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Scene Parsing Cityscapes test VCD No Coarse mIoU 82.3 # 1
Semantic Segmentation GAMUS VCD mIoU 59.70 # 3
Semantic Segmentation NYU Depth v2 VCD+DeepLab (VGG16) Mean IoU 45.3 # 80
Semantic Segmentation NYU Depth v2 VCD+RedNet (ResNet-50) Mean IoU 50.7% # 47
Semantic Segmentation NYU Depth v2 VCD+ACNet (ResNet-50) Mean IoU 51.9% # 36

Methods