THNet: Transferability-Aware Hierarchical Network for Robust Cross-Domain Object Detection

被引:0
作者
Song, Wu [1 ]
Ren, Sheng [1 ]
Tan, Wenxue [1 ]
Wang, Xiping [1 ]
机构
[1] Hunan Univ Arts & Sci, Sch Comp & Elect Engn, Changde 415000, Peoples R China
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Object detection; Feature extraction; Training; Detectors; Adversarial machine learning; Adaptation models; Prototypes; Residual neural networks; Remote sensing; Mathematical models; Cross-domain object detection; hierarchical domain alignment; domain-consistent loss; transferable attention; adversarial learning;
D O I
10.1109/ACCESS.2024.3480351
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has advanced object detection, but generalizing models from source to target domains remains a challenge due to multi-level domain drift and untransferable information. To address this, we propose a transferability-aware hierarchical domain-consistent object detector (THNet), incorporating instance-level, pixel-level, and image-level alignment subnets for robust cross-domain detection. THNet first aligns local foreground-transferable features through pixel-level adversarial learning and foreground-aware attention, then captures global domain-invariant features via image-level subnet with channel-transferable attention. Additionally, a prototype graph convolutional network alleviates instance distribution differences by maximizing inter-class distances and minimizing intra-class distances. A domain-consistent loss harmonizes training for better convergence in multi-level domain alignment. Extensive experiments demonstrate that THNet outperforms state-of-the-art methods on multiple cross-domain datasets, achieving top accuracies of 51.9%, 46.0%, 41.2%, and 51.9% across different tasks.
引用
收藏
页码:155469 / 155484
页数:16
相关论文
共 41 条
  • [11] Fast R-CNN
    Girshick, Ross
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1440 - 1448
  • [12] Rich feature hierarchies for accurate object detection and semantic segmentation
    Girshick, Ross
    Donahue, Jeff
    Darrell, Trevor
    Malik, Jitendra
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 580 - 587
  • [13] Guo TT, 2019, IEEE IMAGE PROC, P1660, DOI [10.1109/ICIP.2019.8803104, 10.1109/icip.2019.8803104]
  • [14] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [15] Multi-adversarial Faster-RCNN for Unrestricted Object Detection
    He, Zhenwei
    Zhang, Lei
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 6667 - 6676
  • [16] Hsu HK, 2020, IEEE WINT CONF APPL, P738, DOI [10.1109/WACV45572.2020.9093358, 10.1109/wacv45572.2020.9093358]
  • [17] Cross-Domain Weakly-Supervised Object Detection through Progressive Domain Adaptation
    Inoue, Naoto
    Furuta, Ryosuke
    Yamasaki, Toshihiko
    Aizawa, Kiyoharu
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5001 - 5009
  • [18] Johnson-Roberson Matthew, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P746, DOI 10.1109/ICRA.2017.7989092
  • [19] Diversify and Match: A Domain Adaptive Representation Learning Paradigm for Object Detection
    Kim, Taekyung
    Jeong, Minki
    Kim, Seunghyeon
    Choi, Seokeon
    Kim, Changick
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 12448 - 12457
  • [20] A Review of Domain Adaptation without Target Labels
    Kouw, Wouter M.
    Loog, Marco
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (03) : 766 - 785