Obstacle Avoidance Algorithm via Hierarchical Interaction Deep Reinforcement Learning

被引:0
作者
Ding, Zihao [1 ]
Song, Chunlei [1 ]
Xu, Jianhua [1 ]
机构
[1] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
来源
2022 41ST CHINESE CONTROL CONFERENCE (CCC) | 2022年
关键词
reinforcement learning; moving obstacle avoidance; motion planning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The navigation task in a complex scenario is an essential problem in mobile robot technology. The mobile robot obstacle avoidance algorithm plays a vital role in navigation. In the navigation task, the mobile robot has to select the optimal action under different conditions in real-time. This research proposes a novel obstacle avoidance algorithm based on deep reinforcement learning. The proposed algorithm utilizes interacting with the environment in the simulation to update the decision network. The decision network includes the feature extraction module and the hierarchical interaction module. The feature extraction module can extract and identify the features of dynamic obstacles in the scenario. And the hierarchical interaction module can handle the interaction features between the mobile robot and obstacles. Furthermore, a safety module is applied in the algorithm to guarantee mobile robot collision-free. Finally, the experiment is conducted to evaluate the proposed method in the simulation environment. The experiment result verified the safety and effectiveness of the proposed method and proved that the proposed method could ensure the mobile robot completes the task.
引用
收藏
页码:3680 / 3685
页数:6
相关论文
共 17 条
[1]  
Chen CG, 2019, IEEE INT CONF ROBOT, P6015, DOI [10.1109/icra.2019.8794134, 10.1109/ICRA.2019.8794134]
[2]   Robot Navigation in Crowds by Graph Convolutional Networks With Attention Learned From Human Gaze [J].
Chen, Yuying ;
Liu, Congcong ;
Shi, Bertram E. ;
Liu, Ming .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (02) :2754-2761
[3]  
Everett M, 2018, IEEE INT C INT ROBOT, P3052, DOI 10.1109/IROS.2018.8593871
[4]   The dynamic window approach to collision avoidance [J].
Fox, D ;
Burgard, W ;
Thrun, S .
IEEE ROBOTICS & AUTOMATION MAGAZINE, 1997, 4 (01) :23-33
[5]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1007/978-3-642-24797-2, 10.1162/neco.1997.9.1.1]
[6]   FlexPool: A Distributed Model-Free Deep Reinforcement Learning Algorithm for Joint Passengers and Goods Transportation [J].
Manchella, Kaushik ;
Umrawal, Abhishek K. ;
Aggarwal, Vaneet .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (04) :2035-2047
[7]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[8]  
Sutton RS, 2018, ADAPT COMPUT MACH LE, P1
[9]   Reciprocal Velocity Obstacles for real-time multi-agent navigation [J].
van den Berg, Jur ;
Lin, Ming ;
Manocha, Dinesh .
2008 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS 1-9, 2008, :1928-1935
[10]  
van den Berg J, 2011, SPRINGER TRAC ADV RO, V70, P3