Unsupervised Adaptation Across Domain Shifts by Generating Intermediate Data Representations

被引:131
作者
Gopalan, Raghuraman [1 ]
Li, Ruonan [2 ]
Chellappa, Rama [3 ,4 ]
机构
[1] AT&T Labs Res, Multimedia Technol Res Dept, Middletown, NJ 07748 USA
[2] Harvard Univ, Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[3] Univ Maryland, Dept Elect & Comp Engn, College Pk, MD 20742 USA
[4] Univ Maryland, UMIACS, Ctr Automat Res, College Pk, MD 20742 USA
关键词
Domain adaptation; unsupervised; Grassmann manifold; object recognition; OBJECT RECOGNITION; MANIFOLDS; GEOMETRY;
D O I
10.1109/TPAMI.2013.249
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With unconstrained data acquisition scenarios widely prevalent, the ability to handle changes in data distribution across training and testing data sets becomes important. One way to approach this problem is through domain adaptation, and in this paper we primarily focus on the unsupervised scenario where the labeled source domain training data is accompanied by unlabeled target domain test data. We present a two-stage data-driven approach by generating intermediate data representations that could provide relevant information on the domain shift. Starting with a linear representation of domains in the form of generative subspaces of same dimensions for the source and target domains, we first utilize the underlying geometry of the space of these subspaces, the Grassmann manifold, to obtain a 'shortest' geodesic path between the two domains. We then sample points along the geodesic to obtain intermediate cross-domain data representations, using which a discriminative classifier is learnt to estimate the labels of the target data. We subsequently incorporate non-linear representation of domains by considering a Reproducing Kernel Hilbert Space representation, and a low-dimensional manifold representation using Laplacian Eigenmaps, and also examine other domain adaptation settings such as (i) semi-supervised adaptation where the target domain is partially labeled, and (ii) multi-domain adaptation where there could be more than one domain in source and/or target data sets. Finally, we supplement our adaptation technique with (i) fine-grained reference domains that are created by blending samples from source and target data sets to provide some evidence on the actual domain shift, and (ii) a multi-class boosting analysis to obtain robustness to the choice of algorithm parameters. We evaluate our approach for object recognition problems and report competitive results on two widely used Office and Bing adaptation data sets.
引用
收藏
页码:2288 / 2302
页数:15
相关论文
共 58 条
[1]   Riemannian geometry of Grassmann manifolds with a view on algorithmic computation [J].
Absil, PA ;
Mahony, R ;
Sepulchre, R .
ACTA APPLICANDAE MATHEMATICAE, 2004, 80 (02) :199-220
[2]  
[Anonymous], 2010, Advances in neural information processing systems
[3]  
[Anonymous], 2007, Proceedings of the National Conference on Artificial Intelligence
[4]   Speeded-Up Robust Features (SURF) [J].
Bay, Herbert ;
Ess, Andreas ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) :346-359
[5]   Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection [J].
Belhumeur, PN ;
Hespanha, JP ;
Kriegman, DJ .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) :711-720
[6]   Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[7]   A theory of learning from different domains [J].
Ben-David, Shai ;
Blitzer, John ;
Crammer, Koby ;
Kulesza, Alex ;
Pereira, Fernando ;
Vaughan, Jennifer Wortman .
MACHINE LEARNING, 2010, 79 (1-2) :151-175
[8]  
Benbouzid D, 2012, J MACH LEARN RES, V13, P549
[9]  
Blitzer J., 2006, Advances in Neural Information Processing Systems, V19, P137
[10]  
Blitzer J., 2006, P C EMPIRICAL METHOD, P120, DOI DOI 10.3115/1610075.1610094