Class-Specific Reconstruction Transfer Learning for Visual Recognition Across Domains

被引:86
作者
Wang, Shanshan [1 ]
Zhang, Lei [1 ]
Zuo, Wangmeng [2 ]
Zhang, Bob [3 ]
机构
[1] Chongqing Univ, Sch Microelect & Commun Engn, Chongqing 400044, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
[3] Univ Macau, Dept Comp & Informat Sci, Taipa, Macao, Peoples R China
关键词
Image reconstruction; Adaptation models; Machine learning; Data models; Correlation; Semantics; Learning systems; Transfer learning; cross-domain learning; semi-supervised learning; image classification; ADAPTATION; KERNEL; CLASSIFICATION; ALGORITHM; IMAGE;
D O I
10.1109/TIP.2019.2948480
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Subspace learning and reconstruction have been widely explored in recent transfer learning work. Generally, a specially designed projection and reconstruction transfer functions bridging multiple domains for heterogeneous knowledge sharing are wanted. However, we argue that the existing subspace reconstruction based domain adaptation algorithms neglect the class prior, such that the learned transfer function is biased, especially when data scarcity of some class is encountered. Different from those previous methods, in this article, we propose a novel class-wise reconstruction-based adaptation method called Class-specific Reconstruction Transfer Learning (CRTL), which optimizes a well modeled transfer loss function by fully exploiting intra-class dependency and inter-class independency. The merits of the CRTL are three-fold. 1) Using a class-specific reconstruction matrix to align the source domain with the target domain fully exploits the class prior in modeling the domain distribution consistency, which benefits the cross-domain classification. 2) Furthermore, to keep the intrinsic relationship between data and labels after feature augmentation, a projected Hilbert-Schmidt Independence Criterion (pHSIC), that measures the dependency between data and label, is first proposed in transfer learning community by mapping the data from raw space to RKHS. 3) In addition, by imposing low-rank and sparse constraints on the class-specific reconstruction coefficient matrix, the global and local data structure that contributes to domain correlation can be effectively preserved. Extensive experiments on challenging benchmark datasets demonstrate the superiority of the proposed method over state-of-the-art representation-based domain adaptation methods. The demo code is available in https://github.com/wangshanshanCQU/CRTL.
引用
收藏
页码:2424 / 2438
页数:15
相关论文
共 63 条
  • [1] [Anonymous], 2013, Decaf: A deep convolutional activation feature for generic visual recognition
  • [2] [Anonymous], 2016, ARXIV160701719
  • [3] A SINGULAR VALUE THRESHOLDING ALGORITHM FOR MATRIX COMPLETION
    Cai, Jian-Feng
    Candes, Emmanuel J.
    Shen, Zuowei
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2010, 20 (04) : 1956 - 1982
  • [4] Ding ZM, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3453
  • [5] Visual Event Recognition in Videos by Learning from Web Data
    Duan, Lixin
    Xu, Dong
    Tsang, Ivor W.
    Luo, Jiebo
    [J]. 2010 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2010, : 1959 - 1966
  • [6] Duan Lixin, 2012, P INT C MACH LEARN, P711, DOI DOI 10.48550/ARXIV.1206.4660
  • [7] How Do Humans Sketch Objects?
    Eitz, Mathias
    Hays, James
    Alexa, Marc
    [J]. ACM TRANSACTIONS ON GRAPHICS, 2012, 31 (04):
  • [8] Sparse Subspace Clustering: Algorithm, Theory, and Applications
    Elhamifar, Ehsan
    Vidal, Rene
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (11) : 2765 - 2781
  • [9] Robust Subspace Segmentation with Block-diagonal Prior
    Feng, Jiashi
    Lin, Zhouchen
    Xu, Huan
    Yan, Shuicheng
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 3818 - 3825
  • [10] Unsupervised Visual Domain Adaptation Using Subspace Alignment
    Fernando, Basura
    Habrard, Amaury
    Sebban, Marc
    Tuytelaars, Tinne
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 2960 - 2967