UAV-YOLOv8: A Small-Object-Detection Model Based on Improved YOLOv8 for UAV Aerial Photography Scenarios

被引:368
作者
Wang, Gang [1 ]
Chen, Yanfei [1 ]
An, Pei [1 ]
Hong, Hanyu [1 ]
Hu, Jinghu [1 ]
Huang, Tiange [1 ]
机构
[1] Wuhan Inst Technol, Sch Elect & Informat Engn, Hubei Key Lab Opt Informat & Pattern Recognit, Wuhan 430205, Peoples R China
关键词
UAVs; small-object detection; YOLOv8; WIoU; BiFormer; FasterNet;
D O I
10.3390/s23167190
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Unmanned aerial vehicle (UAV) object detection plays a crucial role in civil, commercial, and military domains. However, the high proportion of small objects in UAV images and the limited platform resources lead to the low accuracy of most of the existing detection models embedded in UAVs, and it is difficult to strike a good balance between detection performance and resource consumption. To alleviate the above problems, we optimize YOLOv8 and propose an object detection model based on UAV aerial photography scenarios, called UAV-YOLOv8. Firstly, Wise-IoU (WIoU) v3 is used as a bounding box regression loss, and a wise gradient allocation strategy makes the model focus more on common-quality samples, thus improving the localization ability of the model. Secondly, an attention mechanism called BiFormer is introduced to optimize the backbone network, which improves the model's attention to critical information. Finally, we design a feature processing module named Focal FasterNet block (FFNB) and propose two new detection scales based on this module, which makes the shallow features and deep features fully integrated. The proposed multiscale feature fusion network substantially increased the detection performance of the model and reduces the missed detection rate of small objects. The experimental results show that our model has fewer parameters compared to the baseline model and has a mean detection accuracy higher than the baseline model by 7.7%. Compared with other mainstream models, the overall performance of our model is much better. The proposed method effectively improves the ability to detect small objects. There is room to optimize the detection effectiveness of our model for small and feature-less objects (such as bicycle-type vehicles), as we will address in subsequent research.
引用
收藏
页数:27
相关论文
共 54 条
[1]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[2]   A survey on deep learning-based identification of plant and crop diseases from UAV-based aerial images [J].
Bouguettaya, Abdelmalek ;
Zarzour, Hafed ;
Kechida, Ahmed ;
Taberkit, Amine Mohammed .
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2023, 26 (02) :1297-1317
[3]   Road Traffic Monitoring from UAV Images Using Deep Learning Networks [J].
Byun, Sungwoo ;
Shin, In-Kyoung ;
Moon, Jucheol ;
Kang, Jiyoung ;
Choi, Sang-Il .
REMOTE SENSING, 2021, 13 (20)
[4]   Cascade R-CNN: Delving into High Quality Object Detection [J].
Cai, Zhaowei ;
Vasconcelos, Nuno .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6154-6162
[5]   Swin-Transformer-Based YOLOv5 for Small-Object Detection in Remote Sensing Images [J].
Cao, Xuan ;
Zhang, Yanwei ;
Lang, Song ;
Gong, Yan .
SENSORS, 2023, 23 (07)
[6]  
Carion N, 2020, Arxiv, DOI [arXiv:2005.12872, DOI 10.48550/ARXIV.2005.12872]
[7]  
Chen J., 2023, P IEEECVF C COMPUTER
[8]   Research on edge intelligent recognition method oriented to transmission line insulator fault detection [J].
Deng, Fangming ;
Xie, Zhongxin ;
Mao, Wei ;
Li, Bing ;
Shan, Yun ;
Wei, Baoquan ;
Zeng, Han .
INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2022, 139
[9]   Adaptive Sparse Convolutional Networks with Global Context Enhancement for Faster Object Detection on Drone Images [J].
Du, Bowei ;
Huang, Yecheng ;
Chen, Jiaxin ;
Huang, Di .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, :13435-13444
[10]   CenterNet: Keypoint Triplets for Object Detection [J].
Duan, Kaiwen ;
Bai, Song ;
Xie, Lingxi ;
Qi, Honggang ;
Huang, Qingming ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6568-6577