DFTI: Dual-Branch Fusion Network Based on Transformer and Inception for Space Noncooperative Objects

被引:0
|
作者
Zhang, Zhao [1 ]
Zhou, Dong [1 ]
Sun, Guanghui [1 ]
Hu, YuHui [1 ]
Deng, Runran [2 ]
机构
[1] Harbin Inst Technol, Dept Control Sci & Engn, Harbin 150001, Peoples R China
[2] Beijing Inst Spacecraft Syst Engn, Beijing 100094, Peoples R China
基金
中国国家自然科学基金;
关键词
Space vehicles; Feature extraction; Image fusion; Transformers; Task analysis; Visualization; Training; Autoencoder network; deep learning; image fusion; space noncooperative object; transformer; VISIBLE IMAGE FUSION;
D O I
10.1109/TIM.2024.3403182
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Due to adverse illumination in space, noncooperative object perception based on multisource image fusion is crucial for on-orbit maintenance and orbital debris removal. In this article, we first propose a dual-branch multiscale feature extraction encoder combining Transformer block (TB) and Inception block (IB) to extract global and local features of visible and infrared images and establish high-dimensional semantic connections. Second, different from the traditional artificial design fusion strategy, we propose a feature fusion module called cross-convolution feature fusion (CCFF) module, which can achieve image feature level fusion. Based on the above, we propose a dual-branch fusion network based on Transformer and Inception (DFTI) for space noncooperative object, which is an image fusion framework based on autoencoder architecture and unsupervised learning. The fusion image can simultaneously retain the color texture details and contour energy information of space noncooperative objects. Finally, we construct a fusion dataset of infrared and visible images for space noncooperative objects (FIV-SNO) and compare DFTI with seven state-of-the-art methods. In addition, object tracking as a follow-up high-level visual task proves the effectiveness of our method. The experimental results demonstrate that compared with other advanced methods, the entropy (EN) and average gradient (AG) of the fusing images using DFTI network are increased by 0.11 and 0.06, respectively. Our method exhibits excellent performance in both quantitative measures and qualitative evaluation.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Dual-branch network based on transformer for texture recognition
    Liu, Yangqi
    Dong, Hao
    Wang, Guodong
    Chen, Chenglizhao
    DIGITAL SIGNAL PROCESSING, 2024, 153
  • [2] HDCTfusion: Hybrid Dual-Branch Network Based on CNN and Transformer for Infrared and Visible Image Fusion
    Wang, Wenqing
    Li, Lingzhou
    Yang, Yifei
    Liu, Han
    Guo, Runyuan
    SENSORS, 2024, 24 (23)
  • [3] A Dual-Branch Fusion Network Based on Reconstructed Transformer for Building Extraction in Remote Sensing Imagery
    Wang, Yitong
    Wang, Shumin
    Dou, Aixia
    SENSORS, 2024, 24 (02)
  • [4] ENSO-Former: spatiotemporal fusion network based on multivariate and dual-branch transformer for ENSO prediction
    Fang, Wei
    Sha, Yu
    CLIMATE DYNAMICS, 2025, 63 (02)
  • [5] Ship Recognition for Complex SAR Images via Dual-Branch Transformer Fusion Network
    Sun, Zhongzhen
    Leng, Xiangguang
    Zhang, Xianghui
    Xiong, Boli
    Ji, Kefeng
    Kuang, Gangyao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21
  • [6] A dual-branch fracture attribute fusion network based on prior knowledge
    Jiang, Wenbin
    Zhang, Dongmei
    Hui, Gang
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
  • [7] Transformer-Based Dual-Branch Multiscale Fusion Network for Pan-Sharpening Remote Sensing Images
    Li, Zixu
    Li, Jinjiang
    Ren, Lu
    Chen, Zheng
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 614 - 632
  • [8] A Dual-Branch Dynamic Graph Convolution Based Adaptive TransFormer Feature Fusion Network for EEG Emotion Recognition
    Sun, Mingyi
    Cui, Weigang
    Yu, Shuyue
    Han, Hongbin
    Hu, Bin
    Li, Yang
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (04) : 2218 - 2228
  • [9] CosineTR: A dual-branch transformer-based network for semantic line detection
    Zhang, Yuqi
    Ma, Bole
    Jin, Luyang
    Yang, Yuancheng
    Tong, Chao
    PATTERN RECOGNITION, 2025, 158
  • [10] Micro-expression Recognition Based on Dual-Branch Swin Transformer Network
    Xie, Zhihua
    Zhao, Chuwei
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, ICIC 2023, PT II, 2023, 14087 : 544 - 554