VizNav: A Modular Off-Policy Deep Reinforcement Learning Framework for Vision-Based Autonomous UAV Navigation in 3D Dynamic Environments

被引:1
作者
Almahamid, Fadi [1 ]
Grolinger, Katarina [1 ]
机构
[1] Western Univ, Dept Elect & Comp Engn, London, ON N6A 5B9, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
reinforcement learning; autonomous navigation; unmanned aerial vehicle; drone; depth map images; off-policy RL; Prioritized Experience Replay; deep learning; visual navigation framework;
D O I
10.3390/drones8050173
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Unmanned aerial vehicles (UAVs) provide benefits through eco-friendliness, cost-effectiveness, and reduction of human risk. Deep reinforcement learning (DRL) is widely used for autonomous UAV navigation; however, current techniques often oversimplify the environment or impose movement restrictions. Additionally, most vision-based systems lack precise depth perception, while range finders provide a limited environmental overview, and LiDAR is energy-intensive. To address these challenges, this paper proposes VizNav, a modular DRL-based framework for autonomous UAV navigation in dynamic 3D environments without imposing conventional mobility constraints. VizNav incorporates the Twin Delayed Deep Deterministic Policy Gradient (TD3) algorithm with Prioritized Experience Replay and Importance Sampling (PER) to improve performance in continuous action spaces and mitigate overestimations. Additionally, VizNav employs depth map images (DMIs) to enhance visual navigation by accurately estimating objects' depth information, thereby improving obstacle avoidance. Empirical results show that VizNav, by leveraging TD3, improves navigation, and the inclusion of PER and DMI further boosts performance. Furthermore, the deployment of VizNav across various experimental settings confirms its flexibility and adaptability. The framework's architecture separates the agent's learning from the training process, facilitating integration with various DRL algorithms, simulation environments, and reward functions. This modularity creates a potential to influence RL simulation in various autonomous navigation systems, including robotics control and autonomous vehicles.
引用
收藏
页数:27
相关论文
共 57 条
  • [1] Drones Chasing Drones: Reinforcement Learning and Deep Search Area Proposal
    Akhloufi, Moulay A.
    Arola, Sebastien
    Bonnet, Alexandre
    [J]. DRONES, 2019, 3 (03) : 1 - 14
  • [2] Autonomous Unmanned Aerial Vehicle navigation using Reinforcement Learning: A systematic review
    AlMahamid, Fadi
    Grolinger, Katarina
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 115
  • [3] Reinforcement Learning Algorithms: An Overview and Classification
    AlMahamid, Fadi
    Grolinger, Katarina
    [J]. 2021 IEEE CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (CCECE), 2021,
  • [4] Andrew W, 2018, IEEE INT C INT ROBOT, P1124, DOI 10.1109/IROS.2018.8593751
  • [5] Andrychowicz M., 2020, P INT C LEARN REPR, P1
  • [6] Autonomous Navigation via Deep Reinforcement Learning for Resource Constraint Edge Nodes Using Transfer Learning
    Anwar, Aqeel
    Raychowdhury, Arijit
    [J]. IEEE ACCESS, 2020, 8 : 26549 - 26560
  • [7] Vision-Based Navigation Techniques for Unmanned Aerial Vehicles: Review and Challenges
    Arafat, Muhammad Yeasir
    Alam, Muhammad Morshed
    Moh, Sangman
    [J]. DRONES, 2023, 7 (02)
  • [8] Framework for Autonomous UAV Navigation and Target Detection in Global-Navigation-Satellite-System-Denied and Visually Degraded Environments
    Boiteau, Sebastien
    Vanegas, Fernando
    Gonzalez, Felipe
    [J]. REMOTE SENSING, 2024, 16 (03)
  • [9] Bouhamed O., 2020, IEEE INT SYMP CIRC S, P1, DOI [10.1109/iscas45731.2020.9181245, DOI 10.1109/iscas45731.2020.9181245]
  • [10] A Generic Spatiotemporal Scheduling for Autonomous UAVs: A Reinforcement Learning-Based Approach
    Bouhamed, Omar
    Ghazzai, Hakim
    Besbes, Hichem
    Massoud, Yehia
    [J]. IEEE OPEN JOURNAL OF VEHICULAR TECHNOLOGY, 2020, 1 : 93 - 106