Efficient detection of eyes on potato tubers using deep-learning for robotic high-throughput sampling

被引:2
作者
Divyanth, L. G. [1 ]
Khanal, Salik Ram [1 ]
Paudel, Achyut [1 ]
Mattupalli, Chakradhar [2 ]
Karkee, Manoj [1 ]
机构
[1] Washington State Univ, Ctr Precis & Automated Agr Syst, Dept Biol Syst Engn, Prosser, WA 99350 USA
[2] Washington State Univ, Mt Vernon Northwestern Washington Res & Extens Ctr, Dept Plant Pathol, Mt Vernon, WA USA
关键词
tissue sampling robot; machine vision; molecular diagnostics; potato pathogens; FTA card; YOLO; VISION;
D O I
10.3389/fpls.2024.1512632
中图分类号
Q94 [植物学];
学科分类号
071001 ;
摘要
Molecular-based detection of pathogens from potato tubers hold promise, but the initial sample extraction process is labor-intensive. Developing a robotic tuber sampling system, equipped with a fast and precise machine vision technique to identify optimal sampling locations on a potato tuber, offers a viable solution. However, detecting sampling locations such as eyes and stolon scar is challenging due to variability in their appearance, size, and shape, along with soil adhering to the tubers. In this study, we addressed these challenges by evaluating various deep-learning-based object detectors, encompassing You Look Only Once (YOLO) variants of YOLOv5, YOLOv6, YOLOv7, YOLOv8, YOLOv9, YOLOv10, and YOLO11, for detecting eyes and stolon scars across a range of diverse potato cultivars. A robust image dataset obtained from tubers of five potato cultivars (three russet skinned, a red skinned, and a purple skinned) was developed as a benchmark for detection of these sampling locations. The mean average precision at an intersection over union threshold of 0.5 (mAP@0.5) ranged from 0.832 and 0.854 with YOLOv5n to 0.903 and 0.914 with YOLOv10l. Among all the tested models, YOLOv10m showed the optimal trade-off between detection accuracy (mAP@0.5 of 0.911) and inference time (92 ms), along with satisfactory generalization performance when cross-validated among the cultivars used in this study. The model benchmarking and inferences of this study provide insights for advancing the development of a robotic potato tuber sampling device.
引用
收藏
页数:13
相关论文
共 44 条
[31]   A Comprehensive Review of YOLO Architectures in Computer Vision: From YOLOv1 to YOLOv8 and YOLO-NAS [J].
Terven, Juan ;
Cordova-Esparza, Diana-Margarita ;
Romero-Gonzalez, Julio-Alejandro .
MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2023, 5 (04) :1680-1716
[32]  
USDA National Agricultural Statistics Service, 2024, Potatoes summary 2023
[33]   Insect Detection and Identification using YOLO Algorithms on Soybean Crop [J].
Verma, Shani ;
Tripathi, Shrivishal ;
Singh, Anurag ;
Ojha, Muneendra ;
Saxena, Ravi R. .
2021 IEEE REGION 10 CONFERENCE (TENCON 2021), 2021, :272-277
[34]   A review on weed detection using ground-based machine vision and image processing techniques [J].
Wang, Aichen ;
Zhang, Wen ;
Wei, Xinhua .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2019, 158 :226-240
[35]  
Wang A, 2024, Arxiv, DOI [arXiv:2405.14458, DOI 10.48550/ARXIV.2405.14458]
[36]  
Wang CY, 2024, Arxiv, DOI arXiv:2402.13616
[37]   YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors [J].
Wang, Chien-Yao ;
Bochkovskiy, Alexey ;
Liao, Hong-Yuan Mark .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :7464-7475
[38]   You Only Learn One Representation: Unified Network for Multiple Tasks [J].
Wang, Chien-Yao ;
Yeh, I-Hau ;
Liao, Hong-Yuan Mark .
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING, 2023, 39 (03) :691-709
[39]   CSPNet: A New Backbone that can Enhance Learning Capability of CNN [J].
Wang, Chien-Yao ;
Liao, Hong-Yuan Mark ;
Wu, Yueh-Hua ;
Chen, Ping-Yang ;
Hsieh, Jun-Wei ;
Yeh, I-Hau .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :1571-1580
[40]   Object detection in agricultural contexts: A multiple resolution benchmark and comparison to human [J].
Wosner, Omer ;
Farjon, Guy ;
Bar-Hillel, Aharon .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2021, 189