Deep Reinforcement Learning for Mapless Navigation of a Hybrid Aerial Underwater Vehicle with Medium Transition

被引:29
作者
Grando, Ricardo B. [1 ]
de Jesus, Junior C. [1 ]
Kich, Victor A. [2 ]
Kolling, Alisson H. [2 ]
Bortoluzzi, Nicolas P. [1 ]
Pinheiro, Pedro M. [1 ]
Neto, Armando A. [3 ]
Drews, Paulo L. J., Jr. [1 ]
机构
[1] Univ Fed Rio Grande FURG, Ctr Ciencias Computacionais, NAUTEC, Rio Grande, RS, Brazil
[2] Univ Fed Santa Maria UFSM, Santa Maria, RS, Brazil
[3] Univ Fed Minas Gerais, Elect Engn Dept, Belo Horizonte, MG, Brazil
来源
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021) | 2021年
关键词
ROBOTICS;
D O I
10.1109/ICRA48506.2021.9561188
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Since the application of Deep Q-Learning to the continuous action domain in Atari-like games, Deep Reinforcement Learning (Deep-RL) techniques for motion control have been qualitatively enhanced. Nowadays, modern Deep-RL can be successfully applied to solve a wide range of complex decision-making tasks for many types of vehicles. Based on this context, in this paper, we propose the use of Deep-RL to perform autonomous mapless navigation for Hybrid Unmanned Aerial Underwater Vehicles (HUAUVs), robots that can operate in both, air or water media. We developed two approaches, one deterministic and the other stochastic. Our system uses the relative localization of the vehicle and simple sparse range data to train the network. We compared our approaches with an adapted version of the BUG2 algorithm for maples navigation of aerial vehicles. Based on experimental results, we can conclude that Deep-RL-based approaches can be successfully used to perform mapless navigation and obstacle avoidance for HUAUVs. Our vehicle accomplished the navigation in two scenarios, being capable to achieve the desired target through both environments, and even outperforming the behavior-based algorithm on the obstacle-avoidance capability.
引用
收藏
页码:1088 / 1094
页数:7
相关论文
共 37 条
[1]  
Neto AA, 2015, IEEE INT CONF ROBOT, P395, DOI 10.1109/ICRA.2015.7139029
[2]  
Barros G. M., 2020, USING SOFT ACTOR CRI, V2010
[3]  
Carlucho I., 2018, OCEANS 2018 MTSIEEE, P1
[4]  
Chen YF, 2017, IEEE INT C INT ROBOT, P1343, DOI 10.1109/IROS.2017.8202312
[5]  
da Rosa RTS, 2018, IEEE INT C INT ROBOT, P7460, DOI 10.1109/IROS.2018.8593556
[6]  
Drews PLJ, 2014, IEEE INT C INT ROBOT, P4637, DOI 10.1109/IROS.2014.6943220
[7]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[8]  
Furrer F, 2016, STUD COMPUT INTELL, V625, P595, DOI 10.1007/978-3-319-26054-9_23
[9]  
Grando RB, 2020, 2020 XVIII LATIN AMERICAN ROBOTICS SYMPOSIUM, 2020 XII BRAZILIAN SYMPOSIUM ON ROBOTICS AND 2020 XI WORKSHOP OF ROBOTICS IN EDUCATION (LARS-SBR-WRE 2020), P335, DOI [10.1364/CLEOPR.2020.C4B_5, 10.1109/lars/sbr/wre51543.2020.9307015]
[10]  
Gu SX, 2016, PR MACH LEARN RES, V48