S2AC: Self-Supervised Attention Correlation Alignment Based on Mahalanobis Distance for Image Recognition

被引:3
作者
Wang, Zhi-Yong [1 ]
Kang, Dae-Ki [2 ]
Zhang, Cui-Ping [1 ]
机构
[1] Weifang Univ Sci & Technol, Blockchain Lab Agr Vegetables, Weifang 262700, Peoples R China
[2] Dongseo Univ, Dept Comp Engn, 47 Jurye Ro, Busan 47011, South Korea
关键词
domain adaptation; CORAL; self-supervised learning;
D O I
10.3390/electronics12214419
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Susceptibility to domain changes for image classification hinders the application and development of deep neural networks. Domain adaptation (DA) makes use of domain-invariant characteristics to improve the performance of a model trained on labeled data from one domain (source domain) on an unlabeled domain (target) with a different data distribution. But existing DA methods simply use pretrained models (e.g., AlexNet, ResNet) for feature extraction, which are convolutional models that are trapped in localized features and fail to acquire long-distance dependencies. Furthermore, many approaches depend too much on pseudo-labels, which can impair adaptation efficiency and lead to unstable and inconsistent results. In this research, we present S(2)AC, a novel approach for unsupervised deep domain adaptation, that makes use of a stacked attention architecture as a feature map extractor. Our method can fuse domain discrepancy with minimizing a linear transformation of the second statistics (covariances) extended by the p-norm, while simultaneously designing pretext tasks on heuristics to improve the generality of the learning representation. In addition, we have developed a new trainable relative position embedding that not only reduces the model parameters but also enhances model accuracy and expedites the training process. To illustrate our method's efficacy and controllability, we designed extensive experiments based on the Office31, Office_Caltech_10, and OfficeHome datasets. To the best of our knowledge, the proposed method is the first attempt at incorporating attention-based networks and self-supervised learning for image domain adaptation, and has shown promising results.
引用
收藏
页数:19
相关论文
共 31 条
  • [21] Trockman A, 2022, Arxiv, DOI [arXiv:2201.09792, DOI 10.48550/ARXIV.2201.09792]
  • [22] Simultaneous Deep Transfer Across Domains and Tasks
    Tzeng, Eric
    Hoffman, Judy
    Darrell, Trevor
    Saenko, Kate
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 4068 - 4076
  • [23] Tzeng E, 2014, Arxiv, DOI [arXiv:1412.3474, DOI 10.48550/ARXIV.1412.3474]
  • [24] Vaswani A, 2017, ADV NEUR IN, V30
  • [25] Deep Hashing Network for Unsupervised Domain Adaptation
    Venkateswara, Hemanth
    Eusebio, Jose
    Chakraborty, Shayok
    Panchanathan, Sethuraman
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5385 - 5394
  • [26] P-Norm Attention Deep CORAL: Extending Correlation Alignment Using Attention and the P-Norm Loss Function
    Wang, Zhi-Yong
    Kang, Dae-Ki
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (11):
  • [27] d-SNE: Domain Adaptation using Stochastic Neighborhood Embedding
    Xu, Xiang
    Zhou, Xiong
    Venkatesan, Ragav
    Swaminathan, Gurumurthy
    Majumder, Orchid
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2492 - 2500
  • [28] Yang J, 2007, P 15 ACM INT C MULT, P188, DOI DOI 10.1145/1291233.1291276
  • [29] S4L: Self-Supervised Semi-Supervised Learning
    Zhai, Xiaohua
    Oliver, Avital
    Kolesnikov, Alexander
    Beyer, Lucas
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1476 - 1485
  • [30] Zhang H, 2019, PR MACH LEARN RES, V97