Road small target detection based on improved YOLO v5 algorithm

被引:1
作者
Song, Cunli [1 ]
Chai, Weiqin [1 ]
Zhang, Xuesong [1 ]
机构
[1] School of Softmare, Dalian Jiaotong University, Dalian
来源
Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Systems Engineering and Electronics | 2024年 / 46卷 / 10期
关键词
context augmentation module (CAM); deformable convolutional networks (DCN); small target detection; you only look once v5 (YOLO v5);
D O I
10.12305/j.issn.1001-506X.2024.10.04
中图分类号
学科分类号
摘要
In order to solve the problems that small targets on traffic roads faces including detection difficulty, low precision, detection failures, a multi-scale feature fusion target detection improvement algorithm based on the YOLO v5 (you only look once v5) algorithm is proposed. Firstly, the small target detection head is added for adapting to the small target size and alleviating the missed detection. Then, deformable convolutional networks V2 (DCN V2) is introduced to improve the model's learning ability for small targets in motion. The context augmentation module (CAM) is introduced to improve the recognition ability of small targets at a long distance. The replacement loss function is used to improve the bounding box' s localization accuracy, and the spatial pyramid pooling and context spatial pyramid convolution_group (SPPCSPC_group) module is also used to improve the sensory field and feature expression ability of the network. The experiment results show that the proposed algorithm achieves an average accuracy of 95. 2% in the category of small targets in the KITTI dataset.compared with the original YOLO v5 algorithm, the overall average accuracy is improved by 2. 7%. For the detection of small targets, the average accuracy is improved by 3. 1% with a better detection effect, which proves the effectiveness of the proposed algorithm for the detection of small targets on roads. © 2024 Chinese Institute of Electronics. All rights reserved.
引用
收藏
页码:3271 / 3278
页数:7
相关论文
共 31 条
[1]  
GIRSHICK R, DONAHUE J, DARRELL T, Et al., Rich feature hierarchies for accurate object detection and semantic seg- mentation, Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 580-587, (2014)
[2]  
GIRSHICK R., Fast R-CNNECD, Proc. of the IEEE International Conference on Computer Vision, pp. 1440-1448, (2015)
[3]  
REN S Q, HE K M, GIRSHICK R, Et al., Faster R-CNN: towards real-time object detection with region proposal networks, IEEE Trans, on Pattern Analysis and Machine Intelligence, 39, 6, pp. 1137-1149, (2016)
[4]  
HE K M, GKIOXARI G, DOLLAR P, Et al., Mask R-CNN, Proc. of the IEEE International Conference on Computer Vision, pp. 2961-2969, (2017)
[5]  
LIU W, ANGUELOV D, ERHAN D, Et al., SSD: single shot multibox detector, Proc. of the 14th European Conference of Computer Vision-ECCV, pp. 21-37, (2016)
[6]  
REDMON J, DIVVALA S, GIRSHICK R, Et al., You only look once: unified, real-time object detection, Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779-788, (2016)
[7]  
REDMON J, FARHADI A., YOL09000: better, faster, stron- ger, Proc. of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263-7271, (2017)
[8]  
GE Z, LIU S T, WANG F, Et al., YOLOX: exceeding YOLO series in 2021
[9]  
LI C Y, LI L L, JIANG H L, Et al., YOLOv6: a single-stage object detection framework for industrial applications
[10]  
WANG C Y, BOCHKOVSKIY A, LIAOH Y M., YOLOv1: trainable bag-of-freebies sets new state-of-the-art for real-time object detectors, Proc. of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464-7475, (2023)