Abstract
Multi-source domain adaptation has received considerable attention due to its effectiveness of leveraging the knowledge from multiple related sources with different distributions to enhance the learning performance. One of the fundamental challenges in multi-source domain adaptation is how to determine the amount of knowledge transferred from each source domain to the target domain. To address this issue, we propose a new algorithm, called Domain-attention Conditional Wasserstein Distance (DCWD), to learn transferred weights for evaluating the relatedness across the source and target domains. In DCWD, we design a new conditional Wasserstein distance objective function by taking the label information into consideration to measure the distance between a given source domain and the target domain. We also develop an attention scheme to compute the transferred weights of different source domains based on their conditional Wasserstein distances to the target domain. After that, the transferred weights can be used to reweight the source data to determine their importance in knowledge transfer. We conduct comprehensive experiments on several real-world data sets, and the results demonstrate the effectiveness and efficiency of the proposed method.
Original language | English |
---|---|
Article number | 44 |
Number of pages | 19 |
Journal | ACM Transactions on Intelligent Systems and Technology |
Volume | 11 |
Issue number | 4 |
Early online date | 31 May 2020 |
DOIs | |
Publication status | Published - 31 Aug 2020 |
Scopus Subject Areas
- Theoretical Computer Science
- Artificial Intelligence
User-Defined Keywords
- Computing methodologies
- Object recognition
- Transfer learning
- Information systems
- Data mining
- Domain adaptation
- multiple sources
- optimal transport
- attention