On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization

被引:3
|
作者
Thomas, Georgi [1 ]
Sahay, Rajat [1 ]
Jahan, Chowdhury Sadman [1 ]
Manjrekar, Mihir [1 ]
Popp, Dan [1 ]
Savakis, Andreas [1 ]
机构
[1] Rochester Inst Technol, Rochester, NY 14623 USA
关键词
domain adaptation; domain generalization; vision transformers; convolutional neural networks;
D O I
10.3390/s23208409
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The standard feature extraction method in domain adaptation has been convolutional neural networks (CNNs). Recently, attention-based transformer models have emerged as effective alternatives for computer vision tasks. In this paper, we benchmark three attention-based architectures, specifically vision transformer (ViT), shifted window transformer (SWIN), and dual attention vision transformer (DAViT), against convolutional architectures ResNet, HRNet and attention-based ConvNext, to assess the performance of different backbones for domain generalization and adaptation. We incorporate these backbone architectures as feature extractors in the source hypothesis transfer (SHOT) framework for UDA. SHOT leverages the knowledge learned in the source domain to align the image features of unlabeled target data in the absence of source domain data, using self-supervised deep feature clustering and self-training. We analyze the generalization and adaptation performance of these models on standard UDA datasets and aerial UDA datasets. In addition, we modernize the training procedure commonly seen in UDA tasks by adding image augmentation techniques to help models generate richer features. Our results show that ConvNext and SWIN offer the best performance, indicating that the attention mechanism is very beneficial for domain generalization and adaptation with both transformer and convolutional architectures. Our ablation study shows that our modernized training recipe, within the SHOT framework, significantly boosts performance on aerial datasets.
引用
收藏
页数:22
相关论文
共 50 条
  • [11] Domain Attention Model for Domain Generalization in Object Detection
    He, Weixiong
    Zheng, Huicheng
    Lai, Jianhuang
    PATTERN RECOGNITION AND COMPUTER VISION (PRCV 2018), PT IV, 2018, 11259 : 27 - 39
  • [12] GENERALIZATION BOUNDS FOR DOMAIN ADAPTATION VIA DOMAIN TRANSFORMATIONS
    Vural, Elif
    2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [13] Domain generalization via geometric adaptation over augmented data
    Atghaei, Ali
    Rahmati, Mohammad
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [14] Domain Generalization and Adaptation Using Low Rank Exemplar SVMs
    Li, Wen
    Xu, Zheng
    Xu, Dong
    Dai, Dengxin
    Van Gool, Luc
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (05) : 1114 - 1127
  • [15] A Broad Study of Pre-training for Domain Generalization and Adaptation
    Kim, Donghyun
    Wang, Kaihong
    Sclaroff, Stan
    Saenko, Kate
    COMPUTER VISION - ECCV 2022, PT XXXIII, 2022, 13693 : 621 - 638
  • [16] Partially-Labeled Domain Generalization via Multi-Dimensional Domain Adaptation
    Ye, Feiyang
    Bao, Jianghan
    Zhang, Yu
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [17] Unifying Domain Adaptation and Domain Generalization for Robust Prediction Across Minority Racial Groups
    Khoshnevisan, Farzaneh
    Chi, Min
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 521 - 537
  • [18] Attention guided partial domain adaptation for interpretable transfer diagnosis of rotating machinery
    Wang, Gang
    Liu, Dongdong
    Xiang, Jiawei
    Cui, Lingli
    ADVANCED ENGINEERING INFORMATICS, 2024, 62
  • [19] Non-parallel text style transfer with domain adaptation and an attention model
    Mingxuan Hu
    Min He
    Applied Intelligence, 2021, 51 : 4609 - 4622
  • [20] Non-parallel text style transfer with domain adaptation and an attention model
    Hu, Mingxuan
    He, Min
    APPLIED INTELLIGENCE, 2021, 51 (07) : 4609 - 4622