Collision-free path planning for a guava-harvesting robot based on recurrent deep reinforcement learning

被引:118
作者
Lin, Guichao [1 ]
Zhu, Lixue [1 ]
Li, Jinhui [2 ]
Zou, Xiangjun [2 ]
Tang, Yunchao [3 ]
机构
[1] Zhongkai Univ Agr & Engn, Sch Mech & Elect Engn, 501 Zhongkai Rd, Guangzhou 510225, Peoples R China
[2] South China Agr Univ, Coll Engn, 483 Wushan Rd, Guangzhou 510642, Peoples R China
[3] Zhongkai Univ Agr & Engn, Sch Urban & Rural Construct, 501 Zhongkai Rd, Guangzhou 510225, Peoples R China
关键词
Collision-free path planning; Reinforcement learning; Deep deterministic policy gradient; Obstacle detection; Harvesting robot; PICKING;
D O I
10.1016/j.compag.2021.106350
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
In unstructured orchard environments, picking a target fruit without colliding with neighboring branches is a significant challenge for guava-harvesting robots. This paper introduces a fast and robust collision-free pathplanning method based on deep reinforcement learning. A recurrent neural network is first adopted to remember and exploit the past states observed by the robot, then a deep deterministic policy gradient algorithm (DDPG) predicts a collision-free path from the states. A simulation environment is developed and its parameters are randomized during the training phase to enable recurrent DDPG to generalize to real-world scenarios. We also introduce an image processing method that uses a deep neural network to detect obstacles and uses many threedimensional line segments to approximate the obstacles. Simulations show that recurrent DDPG only needs 29 ms to plan a collision-free path with a success rate of 90.90%. Field tests show that recurrent DDPG can increase grasp, detachment, and harvest success rates by 19.43%, 9.11%, and 10.97%, respectively, compared to cases where no collision-free path-planning algorithm is implemented. Recurrent DDPG strikes a strong balance between efficiency and robustness and may be suitable for other fruits.
引用
收藏
页数:9
相关论文
共 23 条
[1]  
Andrychowicz Marcin, 2017, NEURAL INFORM PROCES
[2]   Analysis of a motion planning problem for sweet-pepper harvesting in a dense obstacle environment [J].
Bac, C. Wouter ;
Roorda, Tim ;
Reshef, Roi ;
Berman, Sigal ;
Hemming, Jochen ;
van Henten, Eldert J. .
BIOSYSTEMS ENGINEERING, 2016, 146 :85-97
[3]   Performance Evaluation of a Harvesting Robot for Sweet Pepper [J].
Bac, C. Wouter ;
Hemming, Jochen ;
van Tuijl, B. A. J. ;
Barth, Ruud ;
Wais, Ehud ;
van Henten, Eldert J. .
JOURNAL OF FIELD ROBOTICS, 2017, 34 (06) :1123-1139
[4]   RRT-based path planning for an intelligent litchi-picking manipulator [J].
Cao, Xiaoman ;
Zou, Xiangjun ;
Jia, Chunyang ;
Chen, Mingyou ;
Zeng, Zeqin .
COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2019, 156 :105-118
[5]  
Corke P., 2011, Robotics, Vision and Control. Fundamental Algorithms in MATLAB, DOI DOI 10.1007/978-3-642-20144-8
[6]   LSTM: A Search Space Odyssey [J].
Greff, Klaus ;
Srivastava, Rupesh K. ;
Koutnik, Jan ;
Steunebrink, Bas R. ;
Schmidhuber, Juergen .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (10) :2222-2232
[7]  
Gulcehre C, 2016, PR MACH LEARN RES, V48
[8]  
Hausknecht M., 2015, AAAI FALL S SEQ DEC
[9]  
Lillicrap T. P., 2016, P INT C LEARN REPR S, P1
[10]   Color-, depth-, and shape-based 3D fruit detection [J].
Lin, Guichao ;
Tang, Yunchao ;
Zou, Xiangjun ;
Xiong, Juntao ;
Fang, Yamei .
PRECISION AGRICULTURE, 2020, 21 (01) :1-17