Infrared Weak Target Detection Method Based on Cross Connection and Fusion Attention Mechanism

被引:0
|
作者
Li, Hui [1 ]
Li, Zhengzhou [1 ]
Yang, Yuxin [1 ]
Hao, Congyu [1 ]
Liu, Haitao [1 ]
机构
[1] Chongqing Univ, Sch Microelect & Commun Engn, Chongqing 401331, Peoples R China
基金
中国国家自然科学基金;
关键词
Infrared small targets; Object detection; Crossing connections; Attention mechanism; Multiscale fusion;
D O I
10.3788/gzxb20245309.0910002
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Compared with radar and visible light imaging, infrared imaging has its unique advantages and is widely used in medical imaging, , traffic management, , automatic driving in the civil field, as well as early warning and air defence systems and naval defence systems in the military field. It has the advantages of good concealment, anti-interference and all-weather operation. In complex backgrounds, small targets are usually submerged and look weak, lacking key information such as shape, colour and texture, , generating a large number of spurious false alarms. Traditional methods mainly extract shallow features of the target and background, but due to the lack of effective mining and utilisation of deep features, their adaptability in detecting weak targets in complex scenes is poor, and their ability to detect small targets and their adaptability to the scene need to be improved. Aiming at the problems of low detection performance such as weak signals, unclear features and multiple false alarms of infrared weak targets in complex backgrounds, an infrared weak target detection method based on spanning connection and fusion attention mechanism is proposed. The method combines the attention mechanism with residual networks to extract multiple features of small targets and reduce complex background interference; the bidirectional spanning connection structure fuses feature information at lower and higher levels, highlighting the ability to express the features of weak targets; a high-resolution detection layer is added to regroup the a priori frames of the weak targets and enhance the learning ability of the differences between target and background features; and, finally, the real Gaussian distribution model of target and predicted target frames and calculate their similarity, which solves the problem of sensitivity of target loss regression bias caused by IoU measurements and improves the accuracy of loss regression. The algorithm structure consists of four parts: backbone, neck, head and prediction. The input is an infrared image of size 256x256. The CBS module used by Backbone consists of convolution, batch normalisation and activation functions. The C3 module consists of three convolutional layers and x Resuit modules stitched together. In the last layer of the backbone network, the Convolutional Block Attention Module (CBAM) is introduced, which is fused with the residual network, and the C3CSA module for feature extraction is designed, in order to reduce the background interference in complex scenes. SPPF represents a fast spatial pyramid pooling process. Neck employs a Bidirectional Feature Pyramid Network (BIFPN) that spans connections to fuse low-level detail features to high-level features, as well as transferring high-level semantic information from top to bottom to low-level features. It also adds spanning connections to reduce the loss of some weak target information due to deep feature extraction, to achieve the interaction of global and local information, and to highlight the representation and localisation ability of weak targets at different scales. Upsample represents the up-sampling process. The Head design adds a 64x64 high-resolution feature map weak target detection head, which can avoid the large-scale detection head causing the background interference, and finally predicts the location and confidence information of the target. Comparative tests were conducted on publicly available infrared small target datasets, and the experimental results show that the algorithm has the best performance in detecting infrared small and weak targets in a variety of complex backgrounds, and the average accuracy, recall and speed are significantly improved. The average detection accuracy of this paper's algorithm reaches 98.4%, the model size is only 11.9 MB, and the detection speed is as high as 107 frame/s. By comparing the detection performance of various algorithms in PR curves, ROC curves, and complex scenarios, it can be seen that this paper's algorithm has a better accuracy in detecting weak targets in infrared images in complex scenarios, with a low false alarm rate, and it can be deployed in an embedded terminal for real-time processing.
引用
收藏
页数:12
相关论文
共 23 条
  • [1] Analysis of new top-hat transformation and the application for infrared dim small target detection
    Bai, Xiangzhi
    Zhou, Fugen
    [J]. PATTERN RECOGNITION, 2010, 43 (06) : 2145 - 2156
  • [2] One-Stage Cascade Refinement Networks for Infrared Small Target Detection
    Dai, Yimian
    Li, Xiang
    Zhou, Fei
    Qian, Yulei
    Chen, Yaohong
    Yang, Jian
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [3] Max-Mean and Max-Median filters for detection of small-targets
    Deshpande, SD
    Er, MH
    Ronda, V
    Chan, P
    [J]. SIGNAL AND DATA PROCESSING OF SMALL TARGETS 1999, 1999, 3809 : 74 - 83
  • [4] [回丙伟 Hui Bingwei], 2020, [中国科学数据, China Scientific Data], V5, P1
  • [5] JIANG Xinhao, 2022, Infrared and Laser Engineering, V51
  • [6] Dense Nested Attention Network for Infrared Small Target Detection
    Li, Boyang
    Xiao, Chao
    Wang, Longguang
    Wang, Yingqian
    Lin, Zaiping
    Li, Miao
    An, Wei
    Guo, Yulan
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 1745 - 1758
  • [7] Infrared Small Target Detection Based on Fully Convolutional Neural Network and Visual Saliency
    Liu Jun-ming
    Meng Wei-hua
    [J]. ACTA PHOTONICA SINICA, 2020, 49 (07)
  • [8] An infrared small target detection method based on multiscale local homogeneity measure
    Nie, Jinyan
    Qu, Shaocheng
    Wei, Yantao
    Zhang, Liming
    Deng, Lizhen
    [J]. INFRARED PHYSICS & TECHNOLOGY, 2018, 90 : 186 - 194
  • [9] You Only Look Once: Unified, Real-Time Object Detection
    Redmon, Joseph
    Divvala, Santosh
    Girshick, Ross
    Farhadi, Ali
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 779 - 788
  • [10] Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
    Ren, Shaoqing
    He, Kaiming
    Girshick, Ross
    Sun, Jian
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (06) : 1137 - 1149