Positive Anchor Area Merge Algorithm: A Knowledge Distillation Algorithm for Fruit Detection Tasks Based on Yolov8

被引:1
作者
Shi, Xiangqun [1 ,2 ]
Zhang, Xian [1 ]
Su, Yifan [1 ]
Zhang, Xun [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Zhongshan Inst, Zhongshan 528402, Peoples R China
关键词
YOLO; Knowledge engineering; Training; Accuracy; Prediction algorithms; Feature extraction; Deep learning; Adaptation models; Network architecture; Knowledge transfer; object detection; knowledge distillation; anchor assignment; embedded system;
D O I
10.1109/ACCESS.2025.3544361
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the agricultural sector, employing machine vision technology for fruit target detection holds significant research importance and broad application prospects, such as enabling fruit growth monitoring, yield prediction, and fruit sorting. The Yolov8 model, as the latest model in the field of object detection, boasts advantages including high execution efficiency and detection accuracy. However, when it comes to fruit object detection, which means counting and locating target fruits in an image, the performance of the Yolov8 model shows a noticeable decline compared to its performance on the standard COCO dataset. To address this issue, knowledge distillation is a highly versatile method that uses a large teacher model to guide the training of a smaller student model, thereby improving the detection accuracy of the student model. This thesis proposes a Yolov8 knowledge distillation method tailored for fruit recognition tasks, which improves the network through knowledge distillation and implements a knowledge distillation method based on positive anchor area merging to enhance detection accuracy for fruit recognition tasks. On our self-constructed fruit dataset, which contains over 3,000 images for each category, we compared our model with other similar state-of-the-art models in terms of resource consumption and detection accuracy. While maintaining a low resource overhead, our model achieved an mAP(50) of 99.47%, which is higher than other models that range from 99.1% to 99.3%. In the ablation experiments, we also demonstrated the practical significance of dividing the positive sample area. Finally, we deployed the model on an embedded system for real-time detection of on-site images. These experiments illustrate the practicality of our method for recognizing fruits in real-world scenarios.
引用
收藏
页码:34954 / 34968
页数:15
相关论文
共 30 条
[1]  
Ba LJ, 2014, ADV NEUR IN, V27
[2]  
Bargoti Suchet, 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P3626, DOI 10.1109/ICRA.2017.7989417
[3]   High-Resolution Feature Pyramid Network for Small Object Detection on Drone View [J].
Chen, Zhaodong ;
Ji, Hongbing ;
Zhang, Yongquan ;
Zhu, Zhigang ;
Li, Yifan .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) :475-489
[4]   TOOD: Task-aligned One-stage Object Detection [J].
Feng, Chengjian ;
Zhong, Yujie ;
Gao, Yu ;
Scott, Matthew R. ;
Huang, Weilin .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :3490-3499
[5]  
Gao Feng, 2022, 2022 China Automation Congress (CAC), P4273, DOI 10.1109/CAC57257.2022.10055162
[6]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[7]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587
[8]   Sensors and systems for fruit detection and localization: A review [J].
Gongal, A. ;
Amatya, S. ;
Karkee, M. ;
Zhang, Q. ;
Lewis, K. .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2015, 116 :8-19
[9]   A Comprehensive Overhaul of Feature Distillation [J].
Heo, Byeongho ;
Kim, Jeesoo ;
Yun, Sangdoo ;
Park, Hyojin ;
Kwak, Nojun ;
Choi, Jin Young .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1921-1930
[10]  
Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, 10.48550/arXiv.1503.02531, DOI 10.48550/ARXIV.1503.02531]