Design and Experimental Validation of Deep Reinforcement Learning-Based Fast Trajectory Planning and Control for Mobile Robot in Unknown Environment

被引:119
作者
Chai, Runqi [1 ,2 ]
Niu, Hanlin [1 ]
Carrasco, Joaquin [1 ]
Arvin, Farshad [3 ]
Yin, Hujun [1 ]
Lennox, Barry [1 ]
机构
[1] Univ Manchester, Dept Elect & Elect Engn, Manchester M13 9PL, Lancs, England
[2] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
[3] Univ Durham, Dept Comp Sci, Durham DH1 3LE, England
基金
英国工程与自然科学研究理事会;
关键词
Mobile robots; Trajectory; Planning; Collision avoidance; Training; Robot sensing systems; Noise measurement; Deep reinforcement learning (DRL); mobile robot; motion control; noisy prioritized experience replay (PER); optimal motion planning; recurrent neural network; unexpected obstacles; ROBUST; IMPLEMENTATION; VEHICLES; ASTERISK;
D O I
10.1109/TNNLS.2022.3209154
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article is concerned with the problem of planning optimal maneuver trajectories and guiding the mobile robot toward target positions in uncertain environments for exploration purposes. A hierarchical deep learning-based control framework is proposed which consists of an upper level motion planning layer and a lower level waypoint tracking layer. In the motion planning phase, a recurrent deep neural network (RDNN)-based algorithm is adopted to predict the optimal maneuver profiles for the mobile robot. This approach is built upon a recently proposed idea of using deep neural networks (DNNs) to approximate the optimal motion trajectories, which has been validated that a fast approximation performance can be achieved. To further enhance the network prediction performance, a recurrent network model capable of fully exploiting the inherent relationship between preoptimized system state and control pairs is advocated. In the lower level, a deep reinforcement learning (DRL)-based collision-free control algorithm is established to achieve the waypoint tracking task in an uncertain environment (e.g., the existence of unexpected obstacles). Since this approach allows the control policy to directly learn from human demonstration data, the time required by the training process can be significantly reduced. Moreover, a noisy prioritized experience replay (PER) algorithm is proposed to improve the exploring rate of control policy. The effectiveness of applying the proposed deep learning-based control is validated by executing a number of simulation and experimental case studies. The simulation result shows that the proposed DRL method outperforms the vanilla PER algorithm in terms of training speed. Experimental videos are also uploaded, and the corresponding results confirm that the proposed strategy is able to fulfill the autonomous exploration mission with improved motion planning performance, enhanced collision avoidance ability, and less training time.
引用
收藏
页码:5778 / 5792
页数:15
相关论文
共 50 条
  • [21] Mobile Service Robot Path Planning Using Deep Reinforcement Learning
    Kumaar, A. A. Nippun
    Kochuvila, Sreeja
    [J]. IEEE ACCESS, 2023, 11 : 100083 - 100096
  • [22] Online Continual Safe Reinforcement Learning-based Optimal Control of Mobile Robot Formations
    Ganie, Irfan
    Jagannathan, S.
    [J]. 2024 IEEE CONFERENCE ON CONTROL TECHNOLOGY AND APPLICATIONS, CCTA 2024, 2024, : 519 - 524
  • [23] Hybrid Control of Trajectory Planning for Desired Trajectory and Collision Avoidance Based on Optimization Problem for a Wheeled Mobile Robot
    Kobayashi, Masato
    Motoi, Naoki
    [J]. IEEJ JOURNAL OF INDUSTRY APPLICATIONS, 2020, 9 (04) : 331 - 340
  • [24] Pedestrian Trajectory Prediction Based on SOPD-GAN Used for the Trajectory Planning and Motion Control of Mobile Robot
    Li, Hao
    Qian, Dong-Hai
    Liu, Guang-Yin
    Cui, Ze
    Lei, Jing-Tao
    [J]. IEEE ACCESS, 2023, 11 : 131376 - 131393
  • [25] Continuous Control with Deep Reinforcement Learning for Mobile Robot Navigation
    Xiang, Jiaqi
    Li, Qingdong
    Dong, Xiwang
    Ren, Zhang
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 1501 - 1506
  • [26] Real-time deep learning-based position control of a mobile robot
    Top, Ahmet
    Gokbulut, Muammer
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 138
  • [27] Integral reinforcement learning-based approximate minimum time-energy path planning in an unknown environment
    He, Chenyuan
    Wan, Yan
    Gu, Yixin
    Lewis, Frank L.
    [J]. INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2021, 31 (06) : 1905 - 1922
  • [28] Path Planning for Mobile Robot's Continuous Action Space Based on Deep Reinforcement Learning
    Yan, Tingxing
    Zhang, Yong
    Wang, Bin
    [J]. 2018 INTERNATIONAL CONFERENCE ON BIG DATA AND ARTIFICIAL INTELLIGENCE (BDAI 2018), 2018, : 42 - 46
  • [29] Reinforcement Learning-Based Collision Avoidance and Optimal Trajectory Planning in UAV Communication Networks
    Hsu, Yu-Hsin
    Gau, Rung-Hung
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (01) : 306 - 320
  • [30] A Deep Reinforcement Learning-Based Decentralized Hierarchical Motion Control Strategy for Multiple Amphibious Spherical Robot Systems With Tilting Thrusters
    Yin, He
    Guo, Shuxiang
    Li, Ao
    Shi, Liwei
    Liu, Meng
    [J]. IEEE SENSORS JOURNAL, 2024, 24 (01) : 769 - 779