Lightweight and efficient deep learning models for fruit detection in orchards

被引:5
作者
Yang, Xiaoyao [1 ]
Zhao, Wenyang [1 ]
Wang, Yong [1 ]
Yan, Wei Qi [2 ]
Li, Yanqiang [1 ]
机构
[1] Qilu Univ Technol, Inst Automat, Shandong Acad Sci, Jinan 250014, Peoples R China
[2] Auckland Univ Technol, Auckland, New Zealand
关键词
Recognition of apple; Lightweight network; Attention mechanism; Object detection; Deep learning;
D O I
10.1038/s41598-024-76662-w
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
The accurate recognition of apples in complex orchard environments is a fundamental aspect of the operation of automated picking equipment. This paper aims to investigate the influence of dense targets, occlusion, and the natural environment in practical application scenarios. To this end, it constructs a fruit dataset containing different scenarios and proposes a real-time lightweight detection network, ELD(Efficient Lightweight object Detector). The EGSS(Efficient Ghost-shuffle Slim module) module and MCAttention(Mix channel Attention) are proposed as innovative solutions to the problems of feature extraction and classification. The attention mechanism is employed to construct a novel feature extraction network, which effectively utilizes the low-latitude feature information, significantly enhances the fine-grained feature information and gradient flow of the model, and improves the model's anti-interference ability. Eliminate redundant channels with SlimPAN to further compress the network and optimise functionality. The network as a whole employs the Shape-IOU loss function, which considers the influence of the bounding box itself, thereby enhancing the robustness of the model. Finally, the target detection accuracy is enhanced through the transfer of knowledge from the teacher's network through knowledge distillation, while ensuring that the overall network is sufficiently lightweight. The experimental results demonstrate that the ELD network, designed for fruit detection, achieves an accuracy of 87.4%. It has a relatively low number of parameters (4.3x105\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$4.3 \times 10<^>5$$\end{document}), a GLOPs of only 1.7, and a high FPS of 156. This network can achieve high accuracy while consuming fewer computational resources and performing better than other networks.
引用
收藏
页数:20
相关论文
共 61 条
[11]   Lightweight safflower cluster detection based on YOLOv5 [J].
Guo, Hui ;
Wu, Tianlun ;
Gao, Guomin ;
Qiu, Zhaoxin ;
Chen, Haiyang .
SCIENTIFIC REPORTS, 2024, 14 (01)
[12]   GhostNet: More Features from Cheap Operations [J].
Han, Kai ;
Wang, Yunhe ;
Tian, Qi ;
Guo, Jianyuan ;
Xu, Chunjing ;
Xu, Chang .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :1577-1586
[13]  
Hinton G, 2015, Arxiv, DOI [arXiv:1503.02531, DOI 10.48550/ARXIV.1503.02531]
[14]   Coordinate Attention for Efficient Mobile Network Design [J].
Hou, Qibin ;
Zhou, Daquan ;
Feng, Jiashi .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :13708-13717
[15]   Searching for MobileNetV3 [J].
Howard, Andrew ;
Sandler, Mark ;
Chu, Grace ;
Chen, Liang-Chieh ;
Chen, Bo ;
Tan, Mingxing ;
Wang, Weijun ;
Zhu, Yukun ;
Pang, Ruoming ;
Vasudevan, Vijay ;
Le, Quoc V. ;
Adam, Hartwig .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :1314-1324
[16]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[17]  
Jaderberg M, 2015, ADV NEUR IN, V28
[18]   Screening obstructive sleep apnea patients via deep learning of knowledge distillation in the lateral cephalogram [J].
Kim, Min-Jung ;
Jeong, Jiheon ;
Lee, Jung-Wook ;
Kim, In-Hwan ;
Park, Jae-Woo ;
Roh, Jae-Yon ;
Kim, Namkug ;
Kim, Su-Jung .
SCIENTIFIC REPORTS, 2023, 13 (01)
[19]   Instance, Scale, and Teacher Adaptive Knowledge Distillation for Visual Detection in Autonomous Driving [J].
Lan, Qizhen ;
Tian, Qing .
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (03) :2358-2370
[20]  
Lehnert C, 2016, IEEE INT CONF ROBOT, P2428, DOI 10.1109/ICRA.2016.7487394