Transferable Representation Learning with Deep Adaptation Networks

被引:504
作者
Long, Mingsheng [1 ]
Cao, Yue [1 ]
Cao, Zhangjie [1 ]
Wang, Jianmin [1 ]
Jordan, Michael [2 ]
机构
[1] Tsinghua Univ, Sch Software, Beijing 100084, Peoples R China
[2] Univ Calif Berkeley, Dept EECS, Dept Stat, Berkeley, CA 94720 USA
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Task analysis; Kernel; Adaptation models; Convolutional neural networks; Gallium nitride; Testing; Domain adaptation; deep learning; convolutional neural network; two-sample test; multiple kernel learning; KERNEL;
D O I
10.1109/TPAMI.2018.2868685
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Domain adaptation studies learning algorithms that generalize across source domains and target domains that exhibit different distributions. Recent studies reveal that deep neural networks can learn transferable features that generalize well to similar novel tasks. However, as deep features eventually transition from general to specific along the network, feature transferability drops significantly in higher task-specific layers with increasing domain discrepancy. To formally reduce the effects of this discrepancy and enhance feature transferability in task-specific layers, we develop a novel framework for deep adaptation networks that extends deep convolutional neural networks to domain adaptation problems. The framework embeds the deep features of all task-specific layers into reproducing kernel Hilbert spaces (RKHSs) and optimally matches different domain distributions. The deep features are made more transferable by exploiting low-density separation of target-unlabeled data in very deep architectures, while the domain discrepancy is further reduced via the use of multiple kernel learning that enhances the statistical power of kernel embedding matching. The overall framework is cast in a minimax game setting. Extensive empirical evidence shows that the proposed networks yield state-of-the-art results on standard visual domain-adaptation benchmarks.
引用
收藏
页码:3071 / 3085
页数:15
相关论文
共 56 条
[1]  
[Anonymous], 2006, IEEE T NEURAL NETWOR
[2]  
[Anonymous], INT CONF COMPUTAT
[3]  
[Anonymous], 2007, CALTECH 256 OBJECT C
[4]  
[Anonymous], P NIPS WORKSH TRANSF
[5]  
[Anonymous], PROC CVPR IEEE
[6]  
[Anonymous], ADV NEURAL INFORM PR
[7]  
[Anonymous], 2009, P C LEARN THEOR
[8]  
[Anonymous], 2015, PROCIEEE CONFCOMPUT, DOI DOI 10.1109/CVPR.2015.7298594
[9]   A theory of learning from different domains [J].
Ben-David, Shai ;
Blitzer, John ;
Crammer, Koby ;
Kulesza, Alex ;
Pereira, Fernando ;
Vaughan, Jennifer Wortman .
MACHINE LEARNING, 2010, 79 (1-2) :151-175
[10]  
Ben-David Shai, 2007, NEURIPS