Unsupervised Domain Adaptation via Deep Conditional Adaptation Network

被引:61
作者
Ge, Pengfei [1 ]
Ren, Chuan-Xian [2 ]
Xu, Xiao-Lin [3 ]
Yan, Hong [4 ]
机构
[1] Guangdong Polytech Normal Univ, Sch Math & Syst Sci, Guangzhou 510665, Peoples R China
[2] Sun Yat Sen Univ, Sch Math, Guangzhou 510275, Peoples R China
[3] Guangdong Univ Finance & Econ, Sch Stat & Math, Guangzhou 510320, Peoples R China
[4] City Univ Hong Kong, Dept Elect Engn, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Domain adaptation; Feature extraction; Conditional maximum mean discrepancy; Kernel method;
D O I
10.1016/j.patcog.2022.109088
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised domain adaptation (UDA) aims to generalize the supervised model trained on a source domain to an unlabeled target domain. Previous works mainly rely on the marginal distribution alignment of feature spaces, which ignore the conditional dependence between features and labels, and may suffer from negative transfer. To address this problem, some UDA methods focus on aligning the conditional distributions of feature spaces. However, most of these methods rely on class-specific Maximum Mean Discrepancy or adversarial training, which may suffer from mode collapse and training instability. In this paper, we propose a Deep Conditional Adaptation Network (DCAN) that aligns the conditional distributions by minimizing Conditional Maximum Mean Discrepancy, and extracts discriminant information from the target domain by maximizing the mutual information between samples and the prediction labels. Conditional Maximum Mean Discrepancy measures the difference between conditional distributions directly through their conditional embedding in Reproducing Kernel Hilbert Space, thus DCAN can be trained stably and converge fast. Mutual information can be expressed as the difference between the entropy and conditional entropy of the predicted category variable, thus DCAN can extract the discriminant information of individual and overall distributions in the target domain, simultaneously. In addition, DCAN can be used to address a special scenario, Partial UDA, where the target domain category is a subset of the source domain category. Experiments on both UDA and Partial UDA show that DCAN achieves superior classification performance over state-of-the-art methods.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:14
相关论文
共 56 条
[1]  
[Anonymous], 2011, NEURIPS WORKSHOPS
[2]   JOINT MEASURES AND CROSS-COVARIANCE OPERATORS [J].
BAKER, CR .
TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY, 1973, 186 (459) :273-289
[3]  
Bottou L., 1998, On-line learning in neural networks, V17, P142, DOI DOI 10.1017/CBO9780511569920.003
[4]   Self-Supervised Learning Across Domains [J].
Bucci, Silvia ;
D'Innocente, Antonio ;
Liao, Yujun ;
Carlucci, Fabio Maria ;
Caputo, Barbara ;
Tommasi, Tatiana .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) :5516-5528
[5]   Partial Adversarial Domain Adaptation [J].
Cao, Zhangjie ;
Ma, Lijia ;
Long, Mingsheng ;
Wang, Jianmin .
COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 :139-155
[6]   Unsupervised neural domain adaptation for document image binarization [J].
Castellanos, Francisco J. ;
Gallego, Antonio-Javier ;
Calvo-Zaragoza, Jorge .
PATTERN RECOGNITION, 2021, 119
[7]   Generative attention adversarial classification network for unsupervised domain adaptation [J].
Chen, Wendong ;
Hu, Haifeng .
PATTERN RECOGNITION, 2020, 107
[8]  
Chen Z., 2020, IEEE C COMP VIS PATT, P12706
[9]  
Denker J.S., 1988, P ADV NEUR INF PROC, P323
[10]   Deep feature learning with relative distance comparison for person re-identification [J].
Ding, Shengyong ;
Lin, Liang ;
Wang, Guangrun ;
Chao, Hongyang .
PATTERN RECOGNITION, 2015, 48 (10) :2993-3003