Deep Reinforcement Learning of Map-Based Obstacle Avoidance for Mobile Robot Navigation

被引:1
作者
Chen G. [1 ]
Pan L. [1 ]
Chen Y. [1 ]
Xu P. [2 ]
Wang Z. [1 ]
Wu P. [1 ]
Ji J. [1 ]
Chen X. [1 ]
机构
[1] School of Computer Science and Technology, University of Science and Technology of China, Anhui, Hefei
[2] School of Data Science, University of Science and Technology of China, Anhui, Hefei
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Grid map; Obstacle avoidance; Robot navigation;
D O I
10.1007/s42979-021-00817-z
中图分类号
学科分类号
摘要
Autonomous and safe navigation in complex environments without collisions is particularly important for mobile robots. In this paper, we propose an end-to-end deep reinforcement learning method for mobile robot navigation with map-based obstacle avoidance. Using the experience collected in the simulation environment, a convolutional neural network is trained to predict the proper steering operation of the robot based on its egocentric local grid maps, which can accommodate various sensors and fusion algorithms. We use dueling double DQN with prioritized experienced replay technology to update parameters of the network and integrate curriculum learning techniques to enhance its performance. The trained deep neural network is then transferred and executed on a real-world mobile robot to guide it to avoid local obstacles for long-range navigation. The qualitative and quantitative evaluations of the new approach were performed in simulations and real robot experiments. The results show that the end-to-end map-based obstacle avoidance model is easy to deploy, without any fine-tuning, robust to sensor noise, compatible with different sensors, and better than other related DRL-based models in many evaluation indicators. © 2021, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 38 条
[1]  
Ingrand F., Ghallab M., Deliberation for autonomous robots: a survey, Artif Intell, 247, pp. 10-44, (2017)
[2]  
Minguez J., Lamiraux F., Laumond J.-P., Motion planning and obstacle avoidance, Handbook of Robotics, pp. 1177-1202, (2016)
[3]  
Mohanan M., Salgoankar A., A survey of robotic motion planning in dynamic environments, Robot Auton Syst, 100, pp. 171-185, (2018)
[4]  
Zhang W., Wei S., Teng Y., Zhang J., Wang X., Yan Z., Dynamic obstacle avoidance for unmanned underwater vehicles based on an improved velocity obstacle method, Sensors, 17, 12, (2017)
[5]  
Zhou D., Wang Z., Bandyopadhyay S., Schwager M., Fast, on-line collision avoidance for dynamic vehicles using buffered Voronoi cells, IEEE Robot Autom Lett, 2, 2, pp. 1047-1054, (2017)
[6]  
Rosmann C., Hoffmann F., Bertram T., Integrated online trajectory planning and optimization in distinctive topologies, Robot Auton Syst, 88, pp. 142-153, (2017)
[7]  
Kahn G., Villaflor A., Ding B., Abbeel P., Levine S., Self-supervised deep reinforcement learning with generalized computation graphs for robot navigation, Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), pp. 1-8, (2018)
[8]  
Silver D., Schrittwieser J., Simonyan K., Antonoglou I., Huang A., Guez A., Hubert T., Baker L., Lai M., Bolton A., Et al., Mastering the game of go without human knowledge, Nature, 550, 7676, pp. 354-359, (2017)
[9]  
Vinyals O., Babuschkin I., Czarnecki W.M., Mathieu M., Dudzik A., Chung J., Choi D.H., Powell R., Ewalds T., Georgiev P., Et al., Grandmaster level in starcraft II using multi-agent reinforcement learning, Nature, 575, 7782, pp. 350-354, (2019)
[10]  
Levine S., Pastor P., Krizhevsky A., Ibarz J., Quillen D., Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection, Int J Robot Res, 37, 4-5, pp. 421-436, (2018)