Design and Experiment of a Visual Detection System for Zanthoxylum-Harvesting Robot Based on Improved YOLOv5 Model

被引:6
作者
Guo, Jinkai [1 ,2 ]
Xiao, Xiao [1 ,2 ]
Miao, Jianchi [1 ,2 ]
Tian, Bingquan [1 ,2 ]
Zhao, Jing [1 ,2 ]
Lan, Yubin [3 ]
机构
[1] Shandong Univ Technol, Sch Agr Engn & Food Sci, Zibo 255000, Peoples R China
[2] Shandong Univ Technol, Natl Sub Ctr Int Collaborat Res Precis Agr Aviat P, Zibo 255000, Peoples R China
[3] South China Agr Univ, Coll Elect Engn, Guangzhou 510642, Peoples R China
来源
AGRICULTURE-BASEL | 2023年 / 13卷 / 04期
关键词
YOLOv5; deformable convolution; attention mechanism; visual detection system; Zanthoxylum-harvesting robot;
D O I
10.3390/agriculture13040821
中图分类号
S3 [农学(农艺学)];
学科分类号
0901 ;
摘要
In order to achieve accurate detection of mature Zanthoxylum in their natural environment, a Zanthoxylum detection network based on the YOLOv5 object detection model was proposed. It addresses the issues of irregular shape and occlusion caused by the growth of Zanthoxylum on trees and the overlapping of Zanthoxylum branches and leaves with the fruits, which affect the accuracy of Zanthoxylum detection. To improve the model's generalization ability, data augmentation was performed using different methods. To enhance the directionality of feature extraction and enable the convolution kernel to be adjusted according to the actual shape of each Zanthoxylum cluster, the coordinate attention module and the deformable convolution module were integrated into the YOLOv5 network. Through ablation experiments, the impacts of the attention mechanism and deformable convolution on the performance of YOLOv5 were compared. Comparisons were made using the Faster R-CNN, SSD, and CenterNet algorithms. A Zanthoxylum harvesting robot vision detection platform was built, and the visual detection system was tested. The experimental results showed that using the improved YOLOv5 model, as compared to the original YOLOv5 network, the average detection accuracy for Zanthoxylum in its natural environment was increased by 4.6% and 6.9% in terms of mAP@0.5 and mAP@0.5:0.95, respectively, showing a significant advantage over other network models. At the same time, on the test set of Zanthoxylum with occlusions, the improved model showed increased mAP@0.5 and mAP@0.5:0.95 by 5.4% and 4.7%, respectively, compared to the original model. The improved model was tested on a mobile picking platform, and the results showed that the model was able to accurately identify mature Zanthoxylum in its natural environment at a detection speed of about 89.3 frames per second. This research provides technical support for the visual detection system of intelligent Zanthoxylum-harvesting robots.
引用
收藏
页数:18
相关论文
共 41 条
[1]  
An J., 2019, AGR SCI TECHNOL INF, V6, P57
[2]  
[白强 Bai Qiang], 2022, [农业工程学报, Transactions of the Chinese Society of Agricultural Engineering], V38, P163
[3]  
Bochkovskiy A., 2020, PREPRINT
[4]   Real-Time Detection of Mango Based on Improved YOLOv4 [J].
Cao, Zhipeng ;
Yuan, Ruibo .
ELECTRONICS, 2022, 11 (23)
[5]   MYOLO: A Lightweight Fresh Shiitake Mushroom Detection Model Based on YOLOv3 [J].
Cong, Peichao ;
Feng, Hao ;
Lv, Kunfeng ;
Zhou, Jiachao ;
Li, Shanda .
AGRICULTURE-BASEL, 2023, 13 (02)
[6]   Deformable Convolutional Networks [J].
Dai, Jifeng ;
Qi, Haozhi ;
Xiong, Yuwen ;
Li, Yi ;
Zhang, Guodong ;
Hu, Han ;
Wei, Yichen .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :764-773
[7]   CenterNet: Keypoint Triplets for Object Detection [J].
Duan, Kaiwen ;
Bai, Song ;
Xie, Lingxi ;
Qi, Honggang ;
Huang, Qingming ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6568-6577
[8]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[9]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587
[10]  
Hao J., 2022, J AEROSPACE ENG, V38, P183