Attention-based Dual Supervised Decoder for RGBD Semantic Segmentation

5 Jan 2022  ·  Yang Zhang, Yang Yang, Chenyun Xiong, Guodong Sun, Yanwen Guo ·

Encoder-decoder models have been widely used in RGBD semantic segmentation, and most of them are designed via a two-stream network. In general, jointly reasoning the color and geometric information from RGBD is beneficial for semantic segmentation. However, most existing approaches fail to comprehensively utilize multimodal information in both the encoder and decoder. In this paper, we propose a novel attention-based dual supervised decoder for RGBD semantic segmentation. In the encoder, we design a simple yet effective attention-based multimodal fusion module to extract and fuse deeply multi-level paired complementary information. To learn more robust deep representations and rich multi-modal information, we introduce a dual-branch decoder to effectively leverage the correlations and complementary cues of different tasks. Extensive experiments on NYUDv2 and SUN-RGBD datasets demonstrate that our method achieves superior performance against the state-of-the-art methods.

PDF Abstract

Datasets


Results from the Paper


Ranked #13 on Semantic Segmentation on SUN-RGBD (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Semantic Segmentation NYU Depth v2 AMF (ResNet-50) Mean IoU 52.5% # 30
Semantic Segmentation SUN-RGBD DFormer-L Mean IoU 49.6% # 13

Methods


No methods listed for this paper. Add relevant methods here