Domain Adaptation in the Absence of Source Domain Data

被引:55
作者
Chidlovskii, Boris [1 ]
Clinchant, Stephane [1 ]
Csurka, Gabriela [1 ]
机构
[1] Xerox Res Ctr Europe, 6 Chemin Maupertuis, Meylan, France
来源
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING | 2016年
关键词
Domain adaptation; classification; machine learning; marginalization; emerging applications;
D O I
10.1145/2939672.2939716
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The overwhelming majority of existing domain adaptation methods makes an assumption of freely available source domain data. An equal access to both source and target data makes it possible to measure the discrepancy between their distributions and to build representations common to both target and source domains. In reality, such a simplifying assumption rarely holds, since source data are routinely a subject of legal and contractual constraints between data owners and data customers. When source domain data can not be accessed, decision making procedures are often available for adaptation nevertheless. These procedures are often presented in the form of classification, identification, ranking etc. rules trained on source data and made ready for a direct deployment and later reuse. In other cases, the owner of a source data is allowed to share a few representative examples such as class means. In this paper we address the domain adaptation problem in real world applications, where the reuse of source domain data is limited to classification rules or a few representative examples. We extend the recent techniques of feature corruption and their marginalization, both in supervised and unsupervised settings. We test and compare them on private and publicly available source datasets and show that significant performance gains can be achieved despite the absence of source data and shortage of labeled target data.
引用
收藏
页码:451 / 460
页数:10
相关论文
共 44 条
[1]  
Agrawal R, 2000, SIGMOD REC, V29, P439, DOI 10.1145/335191.335438
[2]  
[Anonymous], 2013, P 31 INT C MACHINE L
[3]  
[Anonymous], 2015, P ICML
[4]  
[Anonymous], 2013, ICML WORKSHOP CHALLE
[5]  
[Anonymous], TKDE
[6]  
[Anonymous], 2013, Advances in Neural Information Processing Systems, DOI DOI 10.48550/ARXIV.1307.1493
[7]  
[Anonymous], 2015, CORR
[8]  
[Anonymous], 2008, P ICML
[9]  
Ben-David S., 2007, NIPS, V19, P137
[10]  
Blitzer R., 2006, P 2006 C EMP METH NA, P120, DOI DOI 10.3115/1610075.1610094