VizNav: A Modular Off-Policy Deep Reinforcement Learning Framework for Vision-Based Autonomous UAV Navigation in 3D Dynamic Environments

被引:2
作者
Almahamid, Fadi [1 ]
Grolinger, Katarina [1 ]
机构
[1] Western Univ, Dept Elect & Comp Engn, London, ON N6A 5B9, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
reinforcement learning; autonomous navigation; unmanned aerial vehicle; drone; depth map images; off-policy RL; Prioritized Experience Replay; deep learning; visual navigation framework;
D O I
10.3390/drones8050173
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Unmanned aerial vehicles (UAVs) provide benefits through eco-friendliness, cost-effectiveness, and reduction of human risk. Deep reinforcement learning (DRL) is widely used for autonomous UAV navigation; however, current techniques often oversimplify the environment or impose movement restrictions. Additionally, most vision-based systems lack precise depth perception, while range finders provide a limited environmental overview, and LiDAR is energy-intensive. To address these challenges, this paper proposes VizNav, a modular DRL-based framework for autonomous UAV navigation in dynamic 3D environments without imposing conventional mobility constraints. VizNav incorporates the Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm with Prioritized Experience Replay and Importance Sampling (PER) to improve performance in continuous action spaces and mitigate overestimations. Additionally, VizNav employs depth map images (DMIs) to enhance visual navigation by accurately estimating objects' depth information, thereby improving obstacle avoidance. Empirical results show that VizNav, by leveraging TD3, improves navigation, and the inclusion of PER and DMI further boosts performance. Furthermore, the deployment of VizNav across various experimental settings confirms its flexibility and adaptability. The framework's architecture separates the agent's learning from the training process, facilitating integration with various DRL algorithms, simulation environments, and reward functions. This modularity creates a potential to influence RL simulation in various autonomous navigation systems, including robotics control and autonomous vehicles.
引用
收藏
页数:27
相关论文
共 57 条
[41]  
Silver D, 2014, PR MACH LEARN RES, V32
[42]   Memory-Based Deep Reinforcement Learning for Obstacle Avoidance in UAV With Limited Environment Knowledge [J].
Singla, Abhik ;
Padakandla, Sindhu ;
Bhatnagar, Shalabh .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (01) :107-118
[43]  
Sutton RS, 2018, ADAPT COMPUT MACH LE, P1
[44]   A Deep Reinforcement Learning Framework for UAV Navigation in Indoor Environments [J].
Walker, Ory ;
Vanegas, Fernando ;
Gonzalez, Felipe ;
Koenig, Sven .
2019 IEEE AEROSPACE CONFERENCE, 2019,
[45]   Deep-Reinforcement-Learning-Based Autonomous UAV Navigation With Sparse Rewards [J].
Wang, Chao ;
Wang, Jian ;
Wang, Jingjing ;
Zhang, Xudong .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (07) :6180-6190
[46]   Autonomous Navigation of UAVs in Large-Scale Complex Environments: A Deep Reinforcement Learning Approach [J].
Wang, Chao ;
Wang, Jian ;
Shen, Yuan ;
Zhang, Xudong .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (03) :2124-2136
[47]  
Wang C, 2017, IEEE GLOB CONF SIG, P858, DOI 10.1109/GlobalSIP.2017.8309082
[48]   Deep-reinforcement-learning-based UAV autonomous navigation and collision avoidance in unknown environments [J].
Wang, Fei ;
Zhu, Xiaoping ;
Zhou, Zhou ;
Tang, Yang .
CHINESE JOURNAL OF AERONAUTICS, 2024, 37 (03) :237-257
[49]  
Yan P, 2020, PROCEEDINGS OF 2020 3RD INTERNATIONAL CONFERENCE ON UNMANNED SYSTEMS (ICUS), P592, DOI 10.1109/ICUS50048.2020.9274899
[50]   Autonomous UAV Navigation in Dynamic Environments with Double Deep Q-Networks [J].
Yang, Yupeng ;
Zhang, Kai ;
Liu, Dahai ;
Song, Houbing .
2020 AIAA/IEEE 39TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC) PROCEEDINGS, 2020,