Instance Adaptive Self-Training for Unsupervised Domain Adaptation

ECCV 2020  ·  Ke Mei, Chuang Zhu, Jiaqi Zou, Shanghang Zhang ·

The divergence between labeled training data and unlabeled testing data is a significant challenge for recent deep learning models. Unsupervised domain adaptation (UDA) attempts to solve such a problem. Recent works show that self-training is a powerful approach to UDA. However, existing methods have difficulty in balancing scalability and performance. In this paper, we propose an instance adaptive self-training framework for UDA on the task of semantic segmentation. To effectively improve the quality of pseudo-labels, we develop a novel pseudo-label generation strategy with an instance adaptive selector. Besides, we propose the region-guided regularization to smooth the pseudo-label region and sharpen the non-pseudo-label region. Our method is so concise and efficient that it is easy to be generalized to other unsupervised domain adaptation methods. Experiments on 'GTA5 to Cityscapes' and 'SYNTHIA to Cityscapes' demonstrate the superior performance of our approach compared with the state-of-the-art methods.

PDF Abstract ECCV 2020 PDF ECCV 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Synthetic-to-Real Translation GTAV-to-Cityscapes Labels IAST mIoU 51.5 # 40
Image-to-Image Translation SYNTHIA-to-Cityscapes IAST(ResNet-101) mIoU (13 classes) 57.0 # 13
Domain Adaptation SYNTHIA-to-Cityscapes IAST (ResNet-101) mIoU 49.8 # 18
Synthetic-to-Real Translation SYNTHIA-to-Cityscapes IAST(ResNet-101) MIoU (13 classes) 57.0 # 21
MIoU (16 classes) 49.8 # 20

Methods


No methods listed for this paper. Add relevant methods here