Inferring Latent Domains for Unsupervised Deep Domain Adaptation

被引:18
作者
Mancini, Massimiliano [1 ,2 ,3 ]
Porzi, Lorenzo [4 ]
Bulo, Samuel Rota [4 ]
Caputo, Barbara [2 ,5 ]
Ricci, Elisa [3 ,6 ]
机构
[1] Sapienza Univ Rome, Dept Comp Control & Management Engn, I-00185 Rome, Italy
[2] Italian Inst Technol, I-10144 Turin, Italy
[3] Fdn Bruno Kessler, Trento, Italy
[4] Mapillary Res, Graz, Austria
[5] Politecn Torino, DAUIN Dept Control & Comp Engn, I-10129 Turin, Italy
[6] Univ Trento, Dept Informat Engn & Comp Sci, I-38122 Trento, Italy
基金
欧洲研究理事会;
关键词
Adaptation models; Data models; Computer architecture; Neural networks; Training; Visualization; Training data; Unsupervised domain adaptation; batch normalization; domain discovery; object recognition;
D O I
10.1109/TPAMI.2019.2933829
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised Domain Adaptation (UDA) refers to the problem of learning a model in a target domain where labeled data are not available by leveraging information from annotated data in a source domain. Most deep UDA approaches operate in a single-source, single-target scenario, i.e., they assume that the source and the target samples arise from a single distribution. However, in practice most datasets can be regarded as mixtures of multiple domains. In these cases, exploiting traditional single-source, single-target methods for learning classification models may lead to poor results. Furthermore, it is often difficult to provide the domain labels for all data points, i.e. latent domains should be automatically discovered. This paper introduces a novel deep architecture which addresses the problem of UDA by automatically discovering latent domains in visual datasets and exploiting this information to learn robust target classifiers. Specifically, our architecture is based on two main components, i.e. a side branch that automatically computes the assignment of each sample to its latent domain and novel layers that exploit domain membership information to appropriately align the distribution of the CNN internal feature representations to a reference distribution. We evaluate our approach on publicly available benchmarks, showing that it outperforms state-of-the-art domain adaptation methods.
引用
收藏
页码:485 / 498
页数:14
相关论文
共 66 条
[31]  
Li D, 2018, AAAI CONF ARTIF INTE, P3490
[32]   Deeper, Broader and Artier Domain Generalization [J].
Li, Da ;
Yang, Yongxin ;
Song, Yi-Zhe ;
Hospedales, Timothy M. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5543-5551
[33]   Domain Generalization and Adaptation Using Low Rank Exemplar SVMs [J].
Li, Wen ;
Xu, Zheng ;
Xu, Dong ;
Dai, Dengxin ;
Van Gool, Luc .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (05) :1114-1127
[34]  
Li Y., 2017, P INT C LEARN REPR W
[35]   Cross-Domain Recognition by Identifying Joint Subspaces of Source Domain and Target Domain [J].
Lin, Yuewei ;
Chen, Jing ;
Cao, Yu ;
Zhou, Youjie ;
Zhang, Lingfeng ;
Tang, Yuan Yan ;
Wang, Song .
IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (04) :1090-1101
[36]  
Long M, 2016, PROCEEDINGS OF SYMPOSIUM OF POLICING DIPLOMACY AND THE BELT & ROAD INITIATIVE, 2016, P136
[37]  
Long MS, 2017, PR MACH LEARN RES, V70
[38]  
Long MS, 2015, PR MACH LEARN RES, V37, P97
[39]   Transfer Sparse Coding for Robust Image Representation [J].
Long, Mingsheng ;
Ding, Guiguang ;
Wang, Jianmin ;
Sun, Jiaguang ;
Guo, Yuchen ;
Yu, Philip S. .
2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, :407-414
[40]   Robust Place Categorization with Deep Domain Generalization [J].
Mancini M. ;
Bulo S.R. ;
Caputo B. ;
Ricci E. .
IEEE Robotics and Automation Letters, 2018, 3 (03) :2093-2100