Navigation method for autonomous mobile robots based on ROS and multi-robot improved Q-learning

被引:0
作者
Hamed, Oussama [1 ]
Hamlich, Mohamed [1 ]
机构
[1] Hassan II Univ, CCPS Lab, ENSAM C, Casablanca, Morocco
关键词
Multi-robot system; Reinforcement learning; Path planning; Robot operating system; Autonomous mobile robot;
D O I
10.1007/s13748-024-00320-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, path planning of multi-autonomous mobile robot systems is one of the interesting topics in scientific research due to its complexity and its wide use in many fields, such as modern industry, war field, and logistics. Q-learning algorithm which is a sort of reinforcement learning is a widely used method in autonomous mobile robot path planning, thanks to its capacity of learning by itself in any environment without the need for prior knowledge. To increase the convergence speed of the Q-learning algorithm and adapt it to robotics and multi-robot systems, the Multi-Robot Improved Q-Learning algorithm (MRIQL) is proposed. The Artificial Potential Field algorithm (APF) is used to initialize the Q-learning. During learning, a restricting mechanism is used to prevent unnecessary actions while exploring. This Improved Q-learning algorithm is adapted to multi-robot system path planning by controlling and adjusting the policies of the robots to generate an optimal and collision-free path for each robot. We introduce a simulation environment for mobile robots based on Robot Operating System (ROS) and Gazebo. The experimental results and the simulation demonstrate the validity and the efficiency of the proposed algorithm.
引用
收藏
页数:9
相关论文
共 13 条
[1]   Development of a hybrid multi-layer control architecture for a cooperative team of N - homogeneous robots [J].
ElAshry, Abdelrahman F. ;
Ramadan, Mohamed M. ;
ElAlaily, Ziad A. ;
Zaied, Mahmoud M. ;
Elias, Catherine M. ;
Shehata, Omar M. ;
Morgan, Elsayed, I .
TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2020, 42 (03) :404-421
[2]  
elsevier, Introduction to Mobile Robot Control, V1st
[3]  
HAMED O, 2021, MENDEL, V27, P23, DOI [10.13164/mendel.2021.2.023, DOI 10.13164/MENDEL.2021.2.023]
[4]  
Hamed O., 2020, Improvised multi-robot cooperation strategy for hunting a dynamic target, DOI [10.1109/ISAECT50560.2020.9523684, DOI 10.1109/ISAECT50560.2020.9523684]
[5]  
Hamed O., 2022, Indonesian Journal of Electrical Engineering and Computer Science, V25, P1, DOI DOI 10.11591/IJEECS.V25.I1.PP159-171
[6]  
IEEE, IEEE Xplore
[7]   Modified Q-learning with distance metric and virtual target on path planning of mobile robot [J].
Low, Ee Soong ;
Ong, Pauline ;
Low, Cheng Yee ;
Omar, Rosli .
EXPERT SYSTEMS WITH APPLICATIONS, 2022, 199
[8]   Solving the optimal path planning of a mobile robot using improved Q-learning [J].
Low, Ee Soong ;
Ong, Pauline ;
Cheah, Kah Chun .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 :143-161
[9]  
Msala Y., 2019, A new robust heterogeneous multi-robot approach based on cloud for task allocation, DOI [10.1109/ICOA.2019.8727618, DOI 10.1109/ICOA.2019.8727618]
[10]  
Takaya K, 2016, INT CONF SYST THEO, P96, DOI 10.1109/ICSTCC.2016.7790647