Polarization-driven Semantic Segmentation via Efficient Attention-bridged Fusion

26 Nov 2020  ·  Kaite Xiang, Kailun Yang, Kaiwei Wang ·

Semantic Segmentation (SS) is promising for outdoor scene perception in safety-critical applications like autonomous vehicles, assisted navigation and so on. However, traditional SS is primarily based on RGB images, which limits the reliability of SS in complex outdoor scenes, where RGB images lack necessary information dimensions to fully perceive unconstrained environments. As preliminary investigation, we examine SS in an unexpected obstacle detection scenario, which demonstrates the necessity of multimodal fusion. Thereby, in this work, we present EAFNet, an Efficient Attention-bridged Fusion Network to exploit complementary information coming from different optical sensors. Specifically, we incorporate polarization sensing to obtain supplementary information, considering its optical characteristics for robust representation of diverse materials. By using a single-shot polarization sensor, we build the first RGB-P dataset which consists of 394 annotated pixel-aligned RGB-Polarization images. A comprehensive variety of experiments shows the effectiveness of EAFNet to fuse polarization and RGB information, as well as the flexibility to be adapted to other sensor combination scenarios.

PDF Abstract

Datasets


Introduced in the Paper:

ZJU-RGB-P

Used in the Paper:

Lost and Found UPLight

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Semantic Segmentation UPLight EAFNet (RGB-AoLP) mIoU 87.53 # 5
Semantic Segmentation UPLight EAFNet (RGB-DoLP) mIoU 87.34 # 6
Semantic Segmentation ZJU-RGB-P EAFNet (RGB-DoLP) mIoU 85.4 # 7
Semantic Segmentation ZJU-RGB-P EAFNet (RGB-AoLP) mIoU 85.7 # 6
Semantic Segmentation ZJU-RGB-P EAFNet (3Path) mIoU 83.4 # 8

Methods


No methods listed for this paper. Add relevant methods here