USD-YOLO: An Enhanced YOLO Algorithm for Small Object Detection in Unmanned Systems Perception

被引:1
作者
Deng, Hongqiang [1 ]
Zhang, Shuzhe [2 ]
Wang, Xiaodong [1 ]
Han, Tianxin [1 ]
Ye, Yun [3 ]
机构
[1] Northeastern Univ, Coll Mech Engn & Automat, Shenyang 110819, Peoples R China
[2] Dalian Maritime Univ, Houston Int Inst, Dalian 116026, Peoples R China
[3] Ningbo Univ, Fac Maritime & Transportat, Ningbo 315211, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2025年 / 15卷 / 07期
基金
中国国家自然科学基金;
关键词
computer vision; object detection; you only look once; deep learning; unmanned system; VEHICLE; RADAR;
D O I
10.3390/app15073795
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In the perception of unmanned systems, small object detection faces numerous challenges, including small size, low resolution, dense distribution, and occlusion, leading to suboptimal perception performance. To address these issues, we propose a specialized algorithm named Unmanned-system Small-object Detection-You Only Look Once (USD-YOLO). First, we designed an innovative module called the Anchor-Free Precision Enhancer to achieve more accurate bounding box overlap measurements and provide a smarter processing mechanism, thereby improving the localization accuracy of candidate boxes for small and densely distributed objects. Second, we introduced the Spatial and Channel Reconstruction Convolution module to reduce redundancy in spatial and channel features while extracting key features of small objects. Additionally, we designed a novel C2f-Global Attention Mechanism module to expand the receptive field and capture more contextual information, optimizing the detection head's ability to handle small and low-resolution objects. We conducted extensive experimental comparisons with state-of-the-art models on three mainstream unmanned system datasets and a real unmanned ground vehicle. The experimental results demonstrate that USD-YOLO achieves higher detection precision and faster speed. On the Citypersons dataset, compared with the baseline, USD-YOLO improves mAP50-95, mAP50, and Recall by 8.5%, 5.9%, and 2.3%, respectively. Additionally, on the Flow-Img and DOTA-v1.0 datasets, USD-YOLO improves mAP50-95 by 2.5% and 2.5%, respectively.
引用
收藏
页数:19
相关论文
共 42 条
[11]   Balanced weighted extreme learning machine for imbalance learning of credit default risk and manufacturing productivity [J].
Khan, Waqar Ahmed .
ANNALS OF OPERATIONS RESEARCH, 2023, 348 (2) :833-861
[12]   SCConv: Spatial and Channel Reconstruction Convolution for Feature Redundancy [J].
Li, Jiafeng ;
Wen, Ying ;
He, Lianghua .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :6153-6162
[13]   Feature Pyramid Networks for Object Detection [J].
Lin, Tsung-Yi ;
Dollar, Piotr ;
Girshick, Ross ;
He, Kaiming ;
Hariharan, Bharath ;
Belongie, Serge .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :936-944
[14]   ZoomInNet: A Novel Small Object Detector in Drone Images with Cross-Scale Knowledge Distillation [J].
Liu, Bi-Yuan ;
Chen, Huai-Xin ;
Huang, Zhou ;
Liu, Xing ;
Yang, Yun-Zhi .
REMOTE SENSING, 2021, 13 (06)
[15]   Path Aggregation Network for Instance Segmentation [J].
Liu, Shu ;
Qi, Lu ;
Qin, Haifang ;
Shi, Jianping ;
Jia, Jiaya .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8759-8768
[16]   SSD: Single Shot MultiBox Detector [J].
Liu, Wei ;
Anguelov, Dragomir ;
Erhan, Dumitru ;
Szegedy, Christian ;
Reed, Scott ;
Fu, Cheng-Yang ;
Berg, Alexander C. .
COMPUTER VISION - ECCV 2016, PT I, 2016, 9905 :21-37
[17]  
Liu Y., 2021, arXiv
[18]   Efficient non-maximum suppression [J].
Neubeck, Alexander ;
Van Gool, Luc .
18TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 3, PROCEEDINGS, 2006, :850-+
[19]  
Ouyang Congjiang, 2022, 2022 IEEE 5th Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), P923, DOI 10.1109/IMCEC55388.2022.10019977
[20]  
Pu Zhengpeng, 2023, 2023 4th International Conference on Computer Engineering and Application (ICCEA), P814, DOI 10.1109/ICCEA58433.2023.10135262