Conditional Adversarial Domain Adaptation

Adversarial learning has been embedded into deep networks to learn disentangled and transferable representations for domain adaptation. Existing adversarial domain adaptation methods may not effectively align different domains of multimodal distributions native in classification problems. In this paper, we present conditional adversarial domain adaptation, a principled framework that conditions the adversarial adaptation models on discriminative information conveyed in the classifier predictions. Conditional domain adversarial networks (CDANs) are designed with two novel conditioning strategies: multilinear conditioning that captures the cross-covariance between feature representations and classifier predictions to improve the discriminability, and entropy conditioning that controls the uncertainty of classifier predictions to guarantee the transferability. With theoretical guarantees and a few lines of codes, the approach has exceeded state-of-the-art results on five datasets.

PDF Abstract NeurIPS 2018 PDF NeurIPS 2018 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Domain Adaptation SVHN-to-MNIST CDAN Accuracy 89.2 # 11
Domain Adaptation USPS-to-MNIST CDAN Accuracy 98.0 # 6
Domain Adaptation VisDA2017 CDAN Accuracy 73.7 # 22

Methods


No methods listed for this paper. Add relevant methods here