EPSNet: Efficient Panoptic Segmentation Network with Cross-layer Attention Fusion

23 Mar 2020  ·  Chia-Yuan Chang, Shuo-En Chang, Pei-Yung Hsiao, Li-Chen Fu ·

Panoptic segmentation is a scene parsing task which unifies semantic segmentation and instance segmentation into one single task. However, the current state-of-the-art studies did not take too much concern on inference time. In this work, we propose an Efficient Panoptic Segmentation Network (EPSNet) to tackle the panoptic segmentation tasks with fast inference speed. Basically, EPSNet generates masks based on simple linear combination of prototype masks and mask coefficients. The light-weight network branches for instance segmentation and semantic segmentation only need to predict mask coefficients and produce masks with the shared prototypes predicted by prototype network branch. Furthermore, to enhance the quality of shared prototypes, we adopt a module called "cross-layer attention fusion module", which aggregates the multi-scale features with attention mechanism helping them capture the long-range dependencies between each other. To validate the proposed work, we have conducted various experiments on the challenging COCO panoptic dataset, which achieve highly promising performance with significantly faster inference speed (53ms on GPU).

PDF Abstract


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Panoptic Segmentation COCO test-dev EPSNet (ResNet-101-FPN) PQ 38.9 # 34
PQst 31.0 # 32
PQth 44.1 # 32