On the Importance of Attention and Augmentations for Hypothesis Transfer in Domain Adaptation and Generalization

被引:3
|
作者
Thomas, Georgi [1 ]
Sahay, Rajat [1 ]
Jahan, Chowdhury Sadman [1 ]
Manjrekar, Mihir [1 ]
Popp, Dan [1 ]
Savakis, Andreas [1 ]
机构
[1] Rochester Inst Technol, Rochester, NY 14623 USA
关键词
domain adaptation; domain generalization; vision transformers; convolutional neural networks;
D O I
10.3390/s23208409
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Unsupervised domain adaptation (UDA) aims to mitigate the performance drop due to the distribution shift between the training and testing datasets. UDA methods have achieved performance gains for models trained on a source domain with labeled data to a target domain with only unlabeled data. The standard feature extraction method in domain adaptation has been convolutional neural networks (CNNs). Recently, attention-based transformer models have emerged as effective alternatives for computer vision tasks. In this paper, we benchmark three attention-based architectures, specifically vision transformer (ViT), shifted window transformer (SWIN), and dual attention vision transformer (DAViT), against convolutional architectures ResNet, HRNet and attention-based ConvNext, to assess the performance of different backbones for domain generalization and adaptation. We incorporate these backbone architectures as feature extractors in the source hypothesis transfer (SHOT) framework for UDA. SHOT leverages the knowledge learned in the source domain to align the image features of unlabeled target data in the absence of source domain data, using self-supervised deep feature clustering and self-training. We analyze the generalization and adaptation performance of these models on standard UDA datasets and aerial UDA datasets. In addition, we modernize the training procedure commonly seen in UDA tasks by adding image augmentation techniques to help models generate richer features. Our results show that ConvNext and SWIN offer the best performance, indicating that the attention mechanism is very beneficial for domain generalization and adaptation with both transformer and convolutional architectures. Our ablation study shows that our modernized training recipe, within the SHOT framework, significantly boosts performance on aerial datasets.
引用
收藏
页数:22
相关论文
共 50 条
  • [21] Unified Deep Supervised Domain Adaptation and Generalization
    Motiian, Saeid
    Piccirilli, Marco
    Adjeroh, Donald A.
    Doretto, Gianfranco
    2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 5716 - 5726
  • [22] MixStyle Neural Networks for Domain Generalization and Adaptation
    Kaiyang Zhou
    Yongxin Yang
    Yu Qiao
    Tao Xiang
    International Journal of Computer Vision, 2024, 132 : 822 - 836
  • [23] On generalization in moment-based domain adaptation
    Werner Zellinger
    Bernhard A. Moser
    Susanne Saminger-Platz
    Annals of Mathematics and Artificial Intelligence, 2021, 89 : 333 - 369
  • [24] Video surveillance using deep transfer learning and deep domain adaptation: Towards better generalization
    Himeur, Yassine
    Al-Maadeed, Somaya
    Kheddar, Hamza
    Al-Maadeed, Noor
    Abualsaud, Khalid
    Mohamed, Amr
    Khattab, Tamer
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 119
  • [25] Domain generalization by marginal transfer learning
    Blanchard, Gilles
    Deshmukh, Aniket Anand
    Dogan, Urun
    Lee, Gyemin
    Scott, Clayton
    Journal of Machine Learning Research, 2021, 22
  • [26] Domain Generalization by Marginal Transfer Learning
    Blanchard, Gilles
    Deshmukh, Aniket Anand
    Dogan, Urun
    Lee, Gyemin
    Scott, Clayton
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [27] Attention guided for partial domain adaptation
    Zhang, Changchun
    Zhao, Qingjie
    INFORMATION SCIENCES, 2021, 547 : 860 - 869
  • [28] Correction to: Non-parallel text style transfer with domain adaptation and an attention model
    Mingxuan Hu
    Min He
    Applied Intelligence, 2021, 51 : 8564 - 8564
  • [29] Improving domain generalization by hybrid domain attention and localized maximum sensitivity
    Ng, Wing W. Y.
    Zhang, Qin
    Zhong, Cankun
    Zhang, Jianjun
    NEURAL NETWORKS, 2024, 171 : 320 - 331
  • [30] Collaborative Contrastive Learning for Hypothesis Domain Adaptation
    Chien, Jen-Tzung
    Yeh, I-Ping
    Mak, Man-Wai
    INTERSPEECH 2024, 2024, : 3225 - 3229