Residential College | false |
Status | 已發表Published |
Guide Subspace Learning for Unsupervised Domain Adaptation | |
Lei Zhang1; Jingru Fu1; Shanshan Wang1; David Zhang2; Zhaoyang Dong3; C. L. Philip Chen4 | |
2020-09-01 | |
Source Publication | IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS |
ISSN | 2162-237X |
Volume | 31Issue:9Pages:3374-3388 |
Abstract | A prevailing problem in many machine learning tasks is that the training (i.e., source domain) and test data (i.e., target domain) have different distribution [i.e., non-independent identical distribution (i.i.d.)]. Unsupervised domain adaptation (UDA) was proposed to learn the unlabeled target data by leveraging the labeled source data. In this article, we propose a guide subspace learning (GSL) method for UDA, in which an invariant, discriminative, and domain-agnostic subspace is learned by three guidance terms through a two-stage progressive training strategy. First, the subspace-guided term reduces the discrepancy between the domains by moving the source closer to the target subspace. Second, the data-guided term uses the coupled projections to map both domains to a unified subspace, where each target sample can be represented by the source samples with a low-rank coefficient matrix that can preserve the global structure of data. In this way, the data from both domains can be well interlaced and the domain-invariant features can be obtained. Third, for improving the discrimination of the subspaces, the label-guided term is constructed for prediction based on source labels and pseudo-target labels. To further improve the model tolerance to label noise, a label relaxation matrix is introduced. For the solver, a two-stage learning strategy with teacher teaches and student feedbacks mode is proposed to obtain the discriminative domain-agnostic subspace. In addition, for handling nonlinear domain shift, a nonlinear GSL (NGSL) framework is formulated with kernel embedding, such that the unified subspace is imposed with nonlinearity. Experiments on various cross-domain visual benchmark databases show that our methods outperform many state-of-the-art UDA methods. The source code is available at https://github.com/Fjr9516/GSL. |
Keyword | Domain Adaptation Subspace Learning Transfer Learning |
DOI | 10.1109/TNNLS.2019.2944455 |
URL | View the original |
Indexed By | SCIE |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering |
WOS Subject | Computer Science, Artificial Intelligence ; Computer Science, Hardware & Architecture ; Computer Science, Theory & Methods ; Engineering, Electrical & Electronic |
WOS ID | WOS:000566342500019 |
Scopus ID | 2-s2.0-85086067029 |
Fulltext Access | |
Citation statistics | |
Document Type | Journal article |
Collection | University of Macau |
Corresponding Author | Lei Zhang |
Affiliation | 1.School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China 2.School of Science and Engineering, Chinese University of Hong Kong (Shenzhen), Shenzhen, 518172, China 3.School of Electrical Engineering and Telecommunications, University of New South Wales, Sydney, 2052, Australia 4.Department of Computer and Information Science, Faculty of Science and Technology, University of Macau, 99999, Macao |
Recommended Citation GB/T 7714 | Lei Zhang,Jingru Fu,Shanshan Wang,et al. Guide Subspace Learning for Unsupervised Domain Adaptation[J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31(9), 3374-3388. |
APA | Lei Zhang., Jingru Fu., Shanshan Wang., David Zhang., Zhaoyang Dong., & C. L. Philip Chen (2020). Guide Subspace Learning for Unsupervised Domain Adaptation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 31(9), 3374-3388. |
MLA | Lei Zhang,et al."Guide Subspace Learning for Unsupervised Domain Adaptation".IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 31.9(2020):3374-3388. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment