Deep Subdomain Adaptation Network for Image Classification

被引:788
作者
Zhu, Yongchun [1 ,2 ]
Zhuang, Fuzhen [1 ,2 ]
Wang, Jindong [3 ]
Ke, Guolin [3 ]
Chen, Jingwu [5 ]
Bian, Jiang [3 ]
Xiong, Hui [4 ]
He, Qing [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[3] Microsoft Res, Beijing, Peoples R China
[4] Rutgers State Univ, New Brunswick, NJ USA
[5] ByteDance, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Adaptation models; Kernel; Feature extraction; Learning systems; Semantics; Training; Domain adaptation; fine grained; subdomain; DOMAIN; KERNEL;
D O I
10.1109/TNNLS.2020.2988928
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a target task where the labeled data are unavailable, domain adaptation can transfer a learner from a different source domain. Previous deep domain adaptation methods mainly learn a global domain shift, i.e., align the global source and target distributions without considering the relationships between two subdomains within the same category of different domains, leading to unsatisfying transfer learning performance without capturing the fine-grained information. Recently, more and more researchers pay attention to subdomain adaptation that focuses on accurately aligning the distributions of the relevant subdomains. However, most of them are adversarial methods that contain several loss functions and converge slowly. Based on this, we present a deep subdomain adaptation network (DSAN) that learns a transfer network by aligning the relevant subdomain distributions of domain-specific layer activations across different domains based on a local maximum mean discrepancy (LMMD). Our DSAN is very simple but effective, which does not need adversarial training and converges fast. The adaptation can be achieved easily with most feedforward network models by extending them with LMMD loss, which can be trained efficiently via backpropagation. Experiments demonstrate that DSAN can achieve remarkable results on both object recognition tasks and digit classification tasks. Our code will be available at https://github.com/easezyc/deep-transfer-learning.
引用
收藏
页码:1713 / 1722
页数:10
相关论文
共 52 条
[1]   A theory of learning from different domains [J].
Ben-David, Shai ;
Blitzer, John ;
Crammer, Koby ;
Kulesza, Alex ;
Pereira, Fernando ;
Vaughan, Jennifer Wortman .
MACHINE LEARNING, 2010, 79 (1-2) :151-175
[2]  
Ben-David Shai, 2007, ADV NEURAL INF PROCE
[3]   Representation Learning: A Review and New Perspectives [J].
Bengio, Yoshua ;
Courville, Aaron ;
Vincent, Pascal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (08) :1798-1828
[4]  
Donahue J, 2014, PR MACH LEARN RES, V32
[5]  
Ganin Y, 2017, ADV COMPUT VIS PATT, P189, DOI 10.1007/978-3-319-58347-1_10
[6]  
Ganin Y, 2015, PR MACH LEARN RES, V37, P1180
[7]   Deep Reconstruction-Classification Networks for Unsupervised Domain Adaptation [J].
Ghifary, Muhammad ;
Kleijn, W. Bastiaan ;
Zhang, Mengjie ;
Balduzzi, David ;
Li, Wen .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :597-613
[8]  
Gong BQ, 2012, PROC CVPR IEEE, P2066, DOI 10.1109/CVPR.2012.6247911
[9]  
Gretton A, 2012, J MACH LEARN RES, V13, P723
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778