DAML: Domain Adaptation Metric Learning

被引:156
作者
Geng, Bo [1 ]
Tao, Dacheng [2 ]
Xu, Chao [1 ]
机构
[1] Peking Univ, Minist Educ, Key Lab Machine Percept, Beijing 100871, Peoples R China
[2] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Quantum Computat & Intelligent Syst, Sydney, NSW 2007, Australia
基金
中国国家自然科学基金;
关键词
Distance metric learning; domain adaptation;
D O I
10.1109/TIP.2011.2134107
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The state-of-the-art metric-learning algorithms cannot perform well for domain adaptation settings, such as cross-domain face recognition, image annotation, etc., because labeled data in the source domain and unlabeled ones in the target domain are drawn from different, but related distributions. In this paper, we propose the domain adaptation metric learning (DAML), by introducing a data-dependent regularization to the conventional metric learning in the reproducing kernel Hilbert space (RKHS). This data-dependent regularization resolves the distribution difference by minimizing the empirical maximum mean discrepancy between source and target domain data in RKHS. Theoretically, by using the empirical Rademacher complexity, we prove risk bounds for the nearest neighbor classifier that uses the metric learned by DAML. Practically, learning the metric in RKHS does not scale up well. Fortunately, we can prove that learning DAML in RKHS is equivalent to learning DAML in the space spanned by principal components of the kernel principle component analysis (KPCA). Thus, we can apply KPCA to select most important principal components to significantly reduce the time cost of DAML. We perform extensive experiments over four well-known face recognition datasets and a large-scale Web image annotation dataset for the cross-domain face recognition and image annotation tasks under various settings, and the results demonstrate the effectiveness of DAML.
引用
收藏
页码:2980 / 2989
页数:10
相关论文
共 42 条
[1]  
[Anonymous], 2002, Principal components analysis
[2]  
[Anonymous], 2009, P 17 ACM INT C MULTI
[3]  
[Anonymous], P IEEE COMP SOC C CO
[4]  
[Anonymous], 2003, P 20 INT C MACH LEAR
[5]  
[Anonymous], P 18 ACM C INF KNOWL
[6]  
[Anonymous], P NAT C ART INT VANC
[7]  
[Anonymous], IEEE T IMAGE PROCESS
[8]  
[Anonymous], ADV NEURAL INF PROCE
[9]  
Balasubramanian M, 2002, SCIENCE, V295
[10]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690