Reinforcement learning-driven dynamic obstacle avoidance for mobile robot trajectory tracking

被引:4
|
作者
Xiao, Hanzhen [1 ]
Chen, Canghao [1 ]
Zhang, Guidong [1 ]
Chen, C. L. Philip [2 ,3 ]
机构
[1] Guangdong Univ Technol, Sch Automat, Guangzhou, Peoples R China
[2] South China Univ Technol, Sch Comp Sci & Engn, Guangzhou, Peoples R China
[3] Pazhou Lab, Ctr Affect Comp & Gen Models, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Obstacle avoidance; Q-Learning; Trajectory tracking; Mobile robot; NAVIGATION;
D O I
10.1016/j.knosys.2024.111974
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we propose a trajectory tracking method based on optimized Q-Learning (QL), which has realtime obstacle avoidance capability, for controlling wheeled mobile robots in dynamic local environments. Based on the observation data and the state of the robot, the designed reinforcement learning (RL) method can determine the obstacle avoidance action during trajectory tracking while simultaneously utilizing controllers to maintain action precision. Through a simple observation space data processing method (OSDPM), the inputting data from the equipped raw lidar is transformed into a dimensionality reduction index vector containing the surrounding environment information of the mobile robot, which can guide QL to quickly correspond the current observation state of the robot to the table state of the QL. To improve the iteration and decision efficiency of the RL method, we optimize the Q -Table structure based on the type of data used. Finally, the simulation results verify the effectiveness of the OSDPM and the obstacle avoidance ability of RL method in unknown local environment.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Moving Target Tracking and Obstacle Avoidance for a Mobile Robot Using MPC
    Eslami, Narges
    Amjadifard, Roya
    2019 27TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE 2019), 2019, : 1163 - 1169
  • [42] Robot target tracking control considering obstacle avoidance based on combination of deep reinforcement learning and PID
    Liu, Yong
    Jiang, Xiao
    Li, Xiang
    Sun, Boxi
    Qian, Sen
    Wu, Yihao
    PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART I-JOURNAL OF SYSTEMS AND CONTROL ENGINEERING, 2025, 239 (03)
  • [43] Obstacle avoidance of a mobile robot using hybrid learning approach
    Er, MJ
    Deng, C
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2005, 52 (03) : 898 - 905
  • [44] Q-Learning for Autonomous Mobile Robot Obstacle Avoidance
    Ribeiro, Tiago
    Goncalves, Fernando
    Garcia, Ines
    Lopes, Gil
    Fernando Ribeiro, A.
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON AUTONOMOUS ROBOT SYSTEMS AND COMPETITIONS (ICARSC 2019), 2019, : 243 - 249
  • [45] Obstacle avoidance path planning for mobile robot based on Ant-Q reinforcement learning algorithm
    Vien, Ngo Anh
    Viet, Nguyen Hoang
    Lee, SeungGwan
    Chung, TaeChoong
    ADVANCES IN NEURAL NETWORKS - ISNN 2007, PT 1, PROCEEDINGS, 2007, 4491 : 704 - +
  • [46] Trajectory Planning of Mobile Robot for Obstacle Avoidance Considering Time and Path Length
    Yorozu, Ayanori
    Yushu, Zou
    Ohya, Akihisa
    INTELLIGENT AUTONOMOUS SYSTEMS 18, VOL 1, IAS18-2023, 2024, 795 : 159 - 174
  • [47] Learning-driven Coarse-to-Fine Articulated Robot Tracking
    Rauch, Christian
    Ivan, Vladimir
    Hospedales, Timothy
    Shotton, Jamie
    Fallon, Maurice
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 6604 - 6610
  • [48] A Comparison of PSO and Reinforcement Learning for Multi-Robot Obstacle Avoidance
    Di Mario, Ezequiel
    Talebpour, Zeynab
    Martinoli, Alcherio
    2013 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2013, : 149 - 156
  • [49] Dynamic Obstacle Avoidance and Path Planning through Reinforcement Learning
    Almazrouei, Khawla
    Kamel, Ibrahim
    Rabie, Tamer
    APPLIED SCIENCES-BASEL, 2023, 13 (14):
  • [50] Dynamic Obstacle Avoidance by Distributed Algorithm based on Reinforcement Learning
    Yaghmaee, F.
    Koohi, H. Reza
    INTERNATIONAL JOURNAL OF ENGINEERING, 2015, 28 (02): : 198 - 204