Research on Robot Navigation Method Integrating Safe Convex Space and Deep Reinforcement Learning

被引:0
|
作者
Dong, Mingze [1 ]
Wen, Zhuanglei [1 ]
Chen, Xiai [1 ]
Yang, Jiongkun [1 ]
Zeng, Tao [1 ]
机构
[1] College of Mechanical and Electrical Engineering, China Jiliang University, Zhejiang, Hangzhou
来源
Binggong Xuebao/Acta Armamentarii | 2024年 / 45卷 / 12期
关键词
deep reinforcement learning; dynamic unknown environment; mobile robot navigation; model predictive control; safe convex space;
D O I
10.12382/bgxb.2023.0982
中图分类号
学科分类号
摘要
A robot navigation method based on deep reinforcement learning (DRL) is proposed for navigating the a robot in the scenario where the global map is unknown and there are dynamic and static obstacles in the environment. Compared to other DRL-based navigation methods applied in complex dynamic environment, the improvements in the designs of action space, state space, and reward function are introduced into the proposed method. Additionally, the proposed method separates the control process from neural network, thus facilitating the simulation research to be effectively implemented in practice. Specifically, the action space is defined by intersecting the safe convex space, calculated from 2D Lidar data, with the kinematic limits of robot. This intersection narrows down the feasible trajectory search space while meeting both short-term dynamic obstacle avoidance and long-term global navigation needs. Reference points are sampled from this action space to form a reference trajectory that the robot follows using a model predictive control (MPC) algorithm. The method also incorporates additional elements such as safe convex space and reference points in the design of state space and reward function. Ablation studies demonstrate the superior navigation success rate, reduced time consumption, and robust generalization capabilities of the proposed method in various static and dynamic environments. © 2024 China Ordnance Industry Corporation. All rights reserved.
引用
收藏
页码:4372 / 4382
页数:10
相关论文
共 25 条
  • [1] HESS W, KOHLER D, RAPP H, Et al., Real-time loop closure in 2D LIDAR SLAM, Proceedings of 2016 IEEE International Conference on Robotics and Automation, pp. 1271-1278, (2016)
  • [2] MUR-ARTAL R, TARD魷S J D., ORB-SLAM 2: an open-source slam system for monocular, stereo, and RGB-D cameras, IEEE Transactions on Robotics, 33, 5, pp. 1255-1262, (2017)
  • [3] P譈TZ S, SIM魷N J S, HERTZBERG J., Move base flex a highly flexible navigation framework for mobile robots, Proceedings of 2018 IEEE / RSJ International Conference on Intelligent Robots and Systems, pp. 3416-3421, (2018)
  • [4] CAI K Q, WANG C Q, CHENG J Y, Et al., Mobile robot path planning in dynamic environments: a survey, Instrumentation, 6, 2, pp. 90-100, (2019)
  • [5] WANG X L, CHEN Y, HU M, Et al., Robot path plannimg for persistent monitoring based on improved deep Q networks [ J], Acta Armamentarii, 45, 6, pp. 1813-1823, (2024)
  • [6] DONG H, YANG J, LI S B, Et al., Research progress of robot motion control based on deep reinforcement learning, Control and Decision, 37, 2, pp. 278-292, (2022)
  • [7] XU X L, CAI P, AHMED Z, Et al., Path planning and dynamic collision avoidance algorithm under COLREGs via deep reinforcement learning, Neurocomputing, 468, pp. 181-197, (2022)
  • [8] YAN N, HUANG S B, KONG C., Reinforcement learning-based autonomous navigation and obstacle avoidance for USVs under partially observable conditions, Mathematical Problems in Engineering, 2021, (2021)
  • [9] PFEIFFER M, SCHAEUBLE M, NIETO J, Et al., From perception to decision: a data-driven approach to end-to-end motion planning for autonomous ground robots, Proceedings of 2017 IEEE International Conference on Robotics and Automation, pp. 1527-1533, (2017)
  • [10] TAI L, PAOLO G, LIU M., Virtual-to-real deep reinforcement learning: Continuous control of mobile robots for mapless navigation[ C ], Proceedings of 2017 IEEE / RSJ International Conference on Intelligent Robots and Systems, pp. 31-36, (2017)