Multimodal target detection method for multi-UAV coordination

被引:0
作者
Sun, Xiaoyong [1 ]
Sun, Bei [1 ]
Guo, Runze [1 ]
Dang, Zhaoyang [1 ]
Zhou, Peida [1 ]
机构
[1] College of Intelligence Science and Technology, National University of Defense Technology, Changsha
来源
Yi Qi Yi Biao Xue Bao/Chinese Journal of Scientific Instrument | 2025年 / 46卷 / 02期
关键词
active perception; multi-UAV coordination; multimodal data; object detection;
D O I
10.19650/j.cnki.cjsi.J2413616
中图分类号
学科分类号
摘要
To solve problems such as limited detection field of view, easy occlusion of target and weak image information of single light source in the current process of single unmanned aerial vehicle (UAV) target detection, and improve the reliable and efficient perception and computing capability of UAV, a multi-modal target detection method for multi-UAV cooperation is proposed in this article. Firstly, a multimodal object detection algorithm based on visible light and infrared fusion is studied, and a dual light fusion model based on a convolutional fusion network driven by visual tasks is proposed. The fused image is fed back to the fusion network through the semantic segmentation network, and a dual light image fusion model with a smaller loss is iteratively trained on the fusion network parameters. Then, the fused image is input into the visual perception enhancement module for image enhancement, eliminating the impact of poor lighting conditions on image quality and improving the preservation of target detail features. The effectiveness of the algorithm is verified on the MSRS dataset. In addition, an active perception process based on distributed biosensing processing is proposed for multi-drone collaborative detection. By calculating the detection confidence of the drone at the location of the sensed target and allocating the detection priority of the host and slave through the release of pheromones, a guidance strategy for multi-drone collaborative detection tasks is completed, achieving target detection in unstructured ground scenes under different lighting conditions. Experimental results show that the algorithm has a 56. 55 ms delay and 45. 84 fps reasoning speed on the unmanned aerial intelligent edge computing platform RK3588, and can accurately detect typical military targets deployed in ground scenes, with an average detection accuracy of 78. 5% . © 2025 Science Press. All rights reserved.
引用
收藏
页码:209 / 220
页数:11
相关论文
共 24 条
[1]  
MA J, LI Y., Application of artificial intelligence in the field of UAV, Radio Engineering, 54, 3, pp. 759-764, (2024)
[2]  
JIA Y N, TIAN SH Y, LI Q., A review of research progress on drone swarms [ J ], Acta Aeronautica et Astronautica Sinica, 41, pp. 4-14, (2020)
[3]  
VIOLA P, JONES M J., Robust real-time face detection[J], International Journal of Computer Vision, 57, pp. 137-154, (2004)
[4]  
DALAL N, TRIGGS B., Histograms of oriented gradients for human detection [ J ], Computer Vision and Pattern Recognition, pp. 886-893, (2005)
[5]  
FELZENSZWALB P F, GIRSHICK P B, MCALLESTER D, Et al., Object detection with discriminatively trained part-based models [ J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 32, 9, pp. 1627-1645, (2010)
[6]  
GIRSHICK R, DONAHUE J, DARRELL T, Et al., Rich feature hierarchies for accurate object detection and semantic segmentation [ J ], 2014 IEEE Conference on Computer Vision and Pattern Recognition, pp. 580-587, (2014)
[7]  
WU X R, QIU T T, WANG Y N., Multi-object detection and segmentation for traffic scene based on improved Mask R-CNN, Chinese Journal of Scientific Instrument, 7, 42, pp. 242-249, (2021)
[8]  
REDMON J, DIVVALA S, GIRSHICK R, Et al., You only look once: Unified, real-time object detection[C], IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788, (2016)
[9]  
FAN Y SH, ZHANG K, NIU L H, Et al., Long-term detection and tracking algorithm for moving vessels by maritime UAVs [ J ], Chinese Journal of Scientific Instrument, 45, 3, pp. 326-335, (2024)
[10]  
LIU W, ANGUELOV D, ERHAN D, Et al., SSD: Single shot multibox detector [ C ], Computer Vision-ECCV, 2016, pp. 21-37, (2016)