A Deep Reinforcement Learning Framework for UAV Navigation in Indoor Environments

被引:37
作者
Walker, Ory [1 ]
Vanegas, Fernando [1 ]
Gonzalez, Felipe [1 ]
Koenig, Sven [2 ]
机构
[1] Queensland Univ Technol, 2 George St, Brisbane, Qld 4000, Australia
[2] Univ Southern Calif, 300 Henry Salvatori Comp Sci Ctr,941 Bloom Walk, Los Angeles, CA 90089 USA
来源
2019 IEEE AEROSPACE CONFERENCE | 2019年
关键词
D O I
10.1109/aero.2019.8742226
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
This paper presents a framework for UAV navigation in indoor environments using a deep reinforcement learning based approach. The implementation models the problem as two seperate problems, a Markov Decision Process (MDP), and a Partially Observable Markov Decision Processes (POMDP), separating the search problem into high-level planning and low-level action under uncertainty. We apply deep learning techniques to this layered problem to produce policies for the framework that allow a UAV to plan, act, and react. The approach is simulated and visualised using Gazebo and is evaluated using policies trained using deep-learning. Using recent deep-learning techniques as the basis of the framework, our results indicate that it is capable of providin smooth navigation for a simulated UAV agent exploring an indoor environment with uncertainty in its position. Once extended to real-world operation, this framework could enable UAVs to be applied in an increasing number of applications, from underground mining and oil refinery surveys and inspections, to search and rescue missions and biosecurity surveys.
引用
收藏
页数:14
相关论文
共 25 条
[1]  
Amato C, 2015, IEEE INT CONF ROBOT, P1241, DOI 10.1109/ICRA.2015.7139350
[2]  
[Anonymous], 2015, CORR
[3]  
[Anonymous], AAAI
[4]  
[Anonymous], 11 PATH FINDER ALGOR
[5]  
Bai HY, 2015, IEEE INT CONF ROBOT, P454, DOI 10.1109/ICRA.2015.7139219
[6]   Algorithms and complexity results for graph-based pursuit evasion [J].
Borie, Richard ;
Tovey, Craig ;
Koenig, Sven .
AUTONOMOUS ROBOTS, 2011, 31 (04) :317-332
[7]  
CARRIO A, 2017, J SENSORS
[8]  
Chen M, 2016, IEEE INT CONF ROBOT, P5427, DOI 10.1109/ICRA.2016.7487754
[9]  
Cheung V, 2016, OPENAI GYM
[10]  
Duan Y, 2016, PR MACH LEARN RES, V48