Guided deep reinforcement learning framework using automated curriculum scheme for accurate motion planning

被引:1
作者
Cho, Deun-Sol [1 ]
Cho, Jae-Min [1 ]
Kim, Won-Tae [2 ]
机构
[1] Koreatech Univ, Major Future Convergence Engn, Cheonan Si 31253, South Korea
[2] Koreatech Univ, Dept Comp Sci Engn, Cheonan Si 31253, South Korea
关键词
Curriculum learning; Deep reinforcement learning; Motion planning; Robotic arm; Unsupervised learning;
D O I
10.1016/j.engappai.2024.109541
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Collaborative robotic arms in smart factories should ensure the safety and interactivity during their operation such as reaching and grasping objects. Especially, the advanced motion planner including the path planning and the motion control functions is essential for human-machine co-working. Since the traditional physics-based motion planning approaches require extreme computational resources to obtain near-optimal solutions, deep reinforcement learning algorithms have been actively adopted and have effectively solved the limitation. They, however, have the easy task preference problem, primarily taking the simpler ways for the more rewards, due to randomly training the agents how to reach the target points in the large-scale search spaces. Therefore, we propose a novel curriculum-based deep reinforcement learning framework that makes the agents learn the motion planning tasks in unbiased ways from the ones with the low complexities to the others with the high complexities. It uses the unsupervised learning algorithms to cluster the target points with the similar task complexities for generating the effective curriculum. In addition, the review and buffer flushing mechanisms are integrated into the framework to mitigate the catastrophic forgetting problem where the agent abruptly lose the previous learned knowledge upon learning new one in the curriculum. The evaluation results of the proposed framework show that the curriculum significantly enhances the success rate on the task with the highest complexity from 12% to 56% and the mechanisms improve the success rate on the tasks with the easier complexities from an average of 66% to 76.5%, despite requiring less training time.
引用
收藏
页数:26
相关论文
共 39 条
[1]   Inside the Virtual Robotics Challenge: Simulating Real-Time Robotic Disaster Response [J].
Agueero, Carlos E. ;
Koenig, Nate ;
Chen, Ian ;
Boyer, Hugo ;
Peters, Steven ;
Hsu, John ;
Gerkey, Brian ;
Paepcke, Steffi ;
Rivero, Jose L. ;
Manzo, Justin ;
Krotkov, Eric ;
Pratt, Gill .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2015, 12 (02) :494-506
[2]  
Altman E., 2021, Constrained markov decision processes
[3]   Adaptive Control and Intersections with Reinforcement Learning [J].
Annaswamy, Anuradha M. .
ANNUAL REVIEW OF CONTROL ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 6 :65-93
[4]   Digital workflow for printability checking and prefabrication in robotic construction 3D printing based on Artificial Intelligence planning [J].
Barjuei, Erfan Shojaei ;
Capitanelli, Alessio ;
Bertolucci, Riccardo ;
Courteille, Eric ;
Mastrogiovanni, Fulvio ;
Maratea, Marco .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
[5]   Deep Deterministic Policy Gradient Virtual Coupling control for thecoordination and manoeuvring of heterogeneous uncertain nonlinearHigh-Speed Trains [J].
Basile, Giacomo ;
Lu, Dario Giuseppe ;
Petrillo, Alberto ;
Santini, Stefania .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
[6]  
Behere S., 2010, A Generic Framework for Robot Motion Planning and Control
[7]  
Coleman D, 2014, Arxiv, DOI [arXiv:1404.3785, 10.48550/arXiv.1404.3785]
[8]   Trajectory planning for a 6-axis robotic arm with particle swarm optimization algorithm [J].
Ekrem, Ozge ;
Aksoy, Bekir .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 122
[9]   Analysis on Deep Reinforcement Learning in Industrial Robotic Arm [J].
Guan, Hengyue .
2020 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND HUMAN-COMPUTER INTERACTION (ICHCI 2020), 2020, :426-430
[10]  
Haarnoja T, 2018, PR MACH LEARN RES, V80