Domain Adaptation

1285 papers with code • 42 benchmarks • 68 datasets

Domain adaptation is the task of adapting models across domains. This is motivated by the challenge where the test and training datasets fall from different data distributions due to some factor. Domain adaptation aims to build machine learning models that can be generalized into a target domain and dealing with the discrepancy across domain distributions.

Further readings:

( Image credit: Unsupervised Image-to-Image Translation Networks )

Libraries

Use these libraries to find Domain Adaptation models and implementations

Most implemented papers

Domain-Adversarial Training of Neural Networks

PaddlePaddle/PaddleSpeech 28 May 2015

Our approach is directly inspired by the theory on domain adaptation suggesting that, for effective domain transfer to be achieved, predictions must be made based on features that cannot discriminate between the training (source) and test (target) domains.

Generalized End-to-End Loss for Speaker Verification

CorentinJ/Real-Time-Voice-Cloning 28 Oct 2017

In this paper, we propose a new loss function called generalized end-to-end (GE2E) loss, which makes the training of speaker verification models more efficient than our previous tuple-based end-to-end (TE2E) loss function.

SuperPoint: Self-Supervised Interest Point Detection and Description

magicleap/SuperGluePretrainedNetwork 20 Dec 2017

This paper presents a self-supervised framework for training interest point detectors and descriptors suitable for a large number of multiple-view geometry problems in computer vision.

Language Models are Few-Shot Learners

openai/gpt-3 NeurIPS 2020

By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do.

Unsupervised Domain Adaptation by Backpropagation

PaddlePaddle/PaddleSpeech 26 Sep 2014

Here, we propose a new approach to domain adaptation in deep architectures that can be trained on large amount of labeled data from the source domain and large amount of unlabeled data from the target domain (no labeled target-domain data is necessary).

Adversarial Discriminative Domain Adaptation

thuml/Transfer-Learning-Library CVPR 2017

Adversarial learning methods are a promising approach to training robust deep networks, and can generate complex samples across diverse domains.

Two at Once: Enhancing Learning and Generalization Capacities via IBN-Net

XingangPan/IBN-Net ECCV 2018

IBN-Net carefully integrates Instance Normalization (IN) and Batch Normalization (BN) as building blocks, and can be wrapped into many advanced deep networks to improve their performances.

Learning to Adapt Structured Output Space for Semantic Segmentation

wasidennis/AdaptSegNet CVPR 2018

In this paper, we propose an adversarial learning method for domain adaptation in the context of semantic segmentation.

Deep CORAL: Correlation Alignment for Deep Domain Adaptation

thuml/Transfer-Learning-Library 6 Jul 2016

CORAL is a "frustratingly easy" unsupervised domain adaptation method that aligns the second-order statistics of the source and target distributions with a linear transformation.

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition

jetpacapp/DeepBeliefSDK 6 Oct 2013

We evaluate whether features extracted from the activation of a deep convolutional network trained in a fully supervised fashion on a large, fixed set of object recognition tasks can be re-purposed to novel generic tasks.