Domain2Vec: Domain Embedding for Unsupervised Domain Adaptation

Xingchao Peng1*     Yichen Li2*     Kate Saenko1    
(*: indicates joint first authors)

1 Boston University     2 Stanford University    

European Conference on Computer Vision (ECCV) 2020

[Paper] [BibTex] [Code (Github)] [Data] [Slides]
Abstract

Conventional unsupervised domain adaptation (UDA) studies the knowledge transfer between a limited number of domains. This neglects the more practical scenario where data are distributed in numerous different domains in the real world. A technique to measure domain similarity is critical for domain adaptation performance. To describe and learn relations between different domains, we propose a novel Domain2Vec model to provide vectorial representations of visual domains based on joint learning of feature disentanglement and Gram matrix. To evaluate the effectiveness of our Domain2Vec model, we create two large-scale cross-domain benchmarks. The first one is TinyDA, which contains 54 domains and about one million MNIST-style images. The second benchmark is DomainBank , which is collected from 56 existing vision datasets. We demonstrate that our embedding is capable of predicting domain similarities that match our intuition about visual relations between different domains. Extensive experiments are conducted to demonstrate the power of our new datasets in benchmarking state-of-the-art multi-source domain adaptation methods, as well as the advantage of our proposed model.

Network Architecture

Figure 2. Network Architecture. Our Domain2Vec architecture achieve deep domain embedding by by joint learning of feature disentanglement and Gram matrix. We employ domain disentanglement (red lines) and class disentanglement (blue lines) to extract domain-specific features and category specific features, both trained adversarially. We further apply a mutual information minimizer to enhance the disentanglement.

Dataset

Figure 4. Examples of our dataset.

Qualitative Results TinyDA

Figure 3. Qualitative Results. we show the (a)t-SNE plot of the embedding result (color indicates different domain) and Domain knowledge graph. The size and color of the circles indicate the number of training examples and the degree of that domain, respectively. The width of the edge shows the domain distance between two domains. as well as the final deep embedding .

Quantitative Results DomainBank

Figure 4. We also show the embeding results for DomainBank.

Quantitative Results

Figure 4. MSDA results on the TinyDA dataset.

Acknowledgements

This work was partially supported by NSF and Honda Research Institute.