L2G: A Simple Local-to-Global Knowledge Transfer Framework for Weakly Supervised Semantic Segmentation

CVPR 2022  ·  Peng-Tao Jiang, YuQi Yang, Qibin Hou, Yunchao Wei ·

Mining precise class-aware attention maps, a.k.a, class activation maps, is essential for weakly supervised semantic segmentation. In this paper, we present L2G, a simple online local-to-global knowledge transfer framework for high-quality object attention mining. We observe that classification models can discover object regions with more details when replacing the input image with its local patches. Taking this into account, we first leverage a local classification network to extract attentions from multiple local patches randomly cropped from the input image. Then, we utilize a global network to learn complementary attention knowledge across multiple local attention maps online. Our framework conducts the global network to learn the captured rich object detail knowledge from a global view and thereby produces high-quality attention maps that can be directly used as pseudo annotations for semantic segmentation networks. Experiments show that our method attains 72.1% and 44.2% mIoU scores on the validation set of PASCAL VOC 2012 and MS COCO 2014, respectively, setting new state-of-the-art records. Code is available at https://github.com/PengtaoJiang/L2G.

PDF Abstract CVPR 2022 PDF CVPR 2022 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Weakly-Supervised Semantic Segmentation COCO 2014 val L2G (DeepLabV2-ResNet101) mIoU 44.2 # 18
Weakly-Supervised Semantic Segmentation PASCAL VOC 2012 test L2G (ResNet101, DeepLab-LargeFOV) Mean IoU 73.0 # 16
Weakly-Supervised Semantic Segmentation PASCAL VOC 2012 val L2G (ResNet101, DeepLab-v2) Mean IoU 72.1 # 18

Methods


No methods listed for this paper. Add relevant methods here