Intelligent Robot in Unknown Environments: Walk Path Using Q-Learning and Deep Q-Learning

被引:1
作者
El Wafi, Mouna [1 ]
Youssefi, My Abdelkader [1 ]
Dakir, Rachid [2 ]
Bakir, Mohamed [1 ]
机构
[1] Hassan First Univ Settat, Fac Sci & Tech, Engn Lab, Ind Management & Innovat, Settat 26000, Morocco
[2] Ibnou Zohr Univ, Polydisciplinary Fac Ouarzazate, Lab Comp Syst & Vis, Ouarzazate 45000, Morocco
来源
AUTOMATION | 2025年 / 6卷 / 01期
关键词
Q-learning; deep Q-learning; reinforcement learning; neural network; path-planning;
D O I
10.3390/automation6010012
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Autonomous navigation is essential for mobile robots to efficiently operate in complex environments. This study investigates Q-learning and Deep Q-learning to improve navigation performance. The research examines their effectiveness in complex maze configurations, focusing on how the epsilon-greedy strategy influences the agent's ability to reach its goal in minimal time using Q-learning. A distinctive aspect of this work is the adaptive tuning of hyperparameters, where alpha and gamma values are dynamically adjusted throughout training. This eliminates the need for manually fixed parameters and enables the learning algorithm to automatically determine optimal values, ensuring adaptability to diverse environments rather than being constrained to specific cases. By integrating neural networks, Deep Q-learning enhances decision-making in complex navigation tasks. Simulations carried out in MATLAB environments validate the proposed approach, illustrating its effectiveness in resource-constrained systems while preserving robust and efficient decision-making. Experimental results demonstrate that adaptive hyperparameter tuning significantly improves learning efficiency, leading to faster convergence and reduced navigation time. Additionally, Deep Q-learning exhibits superior performance in complex environments, showcasing enhanced decision-making capabilities in high-dimensional state spaces. These findings highlight the advantages of reinforcement learning-based navigation and emphasize how adaptive exploration strategies and dynamic parameter adjustments enhance performance across diverse scenarios.
引用
收藏
页数:18
相关论文
共 26 条
[1]  
Aziz I., 2020, Masters Thesis
[2]  
Bahi M, 2018, 2018 3RD INTERNATIONAL CONFERENCE ON PATTERN ANALYSIS AND INTELLIGENT SYSTEMS (PAIS), P268
[3]  
Batina L, 2019, PROCEEDINGS OF THE 28TH USENIX SECURITY SYMPOSIUM, P515
[4]   Study of Q-learning and deep Q-network learning control for a rotary inverted pendulum system [J].
Ben Hazem, Zied .
DISCOVER APPLIED SCIENCES, 2024, 6 (02)
[5]  
Duryea E., 2016, Intell. Control Autom, V7, P129, DOI [10.4236/ica.2016.74012, DOI 10.4236/ICA.2016.74012]
[6]   Q-Learning Algorithms: A Comprehensive Classification and Applications [J].
Jang, Beakcheol ;
Kim, Myeonghwi ;
Harerimana, Gaspard ;
Kim, Jong Wook .
IEEE ACCESS, 2019, 7 :133653-133667
[7]   A Survey of Path Planning Algorithms for Mobile Robots [J].
Karur, Karthik ;
Sharma, Nitin ;
Dharmatti, Chinmay ;
Siegel, Joshua E. .
VEHICLES, 2021, 3 (03) :448-468
[8]   Dynamic Path Planning of Unknown Environment Based on Deep Reinforcement Learning [J].
Lei, Xiaoyun ;
Zhang, Zhian ;
Dong, Peifang .
JOURNAL OF ROBOTICS, 2018, 2018
[9]  
Liu F, 2020, CHIN CONTR CONF, P3730, DOI 10.23919/CCC50068.2020.9188890
[10]   Path planning techniques for mobile robots: Review and prospect [J].
Liu, Lixing ;
Wang, Xu ;
Yang, Xin ;
Liu, Hongjie ;
Li, Jianping ;
Wang, Pengfei .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 227