Drone Deep Reinforcement Learning: A Review

被引:153
作者
Azar, Ahmad Taher [1 ,2 ]
Koubaa, Anis [1 ]
Ali Mohamed, Nada [3 ]
Ibrahim, Habiba A. [4 ]
Ibrahim, Zahra Fathy [3 ]
Kazim, Muhammad [1 ,5 ]
Ammar, Adel [1 ]
Benjdira, Bilel [1 ]
Khamis, Alaa M. [6 ]
Hameed, Ibrahim A. [7 ]
Casalino, Gabriella [8 ]
机构
[1] Prince Sultan Univ, Coll Comp & Informat Sci, Riyadh 11586, Saudi Arabia
[2] Benha Univ, Fac Comp & Artificial Intelligence, Banha 13518, Egypt
[3] Nile Univ Campus, Sch Engn & Appl Sci, Juhayna Sq, Giza 60411, Egypt
[4] Nile Univ, Smart Engn Syst Res Ctr SESC, Giza 12588, Egypt
[5] Harbin Inst Technol, Res Inst Intelligent Control & Syst, Harbin 150080, Peoples R China
[6] Gen Motors Canada, 500 Wentworth St W, Oshawa, ON L1J 6J2, Canada
[7] Norwegian Univ Sci & Technol, Dept ICT & Nat Sci, Larsgardsvegen 2, N-6009 Alesund, Norway
[8] Univ Bari, Dept Informat, I-70125 Bari, Italy
关键词
unmanned aerial vehicles; UAVs; guidance; navigation; control; machine learning; deep reinforcement learning (DRL); literature review; UNMANNED AERIAL VEHICLES; STRUCTURE-FROM-MOTION; UAV; IMAGES;
D O I
10.3390/electronics10090999
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned Aerial Vehicles (UAVs) are increasingly being used in many challenging and diversified applications. These applications belong to the civilian and the military fields. To name a few; infrastructure inspection, traffic patrolling, remote sensing, mapping, surveillance, rescuing humans and animals, environment monitoring, and Intelligence, Surveillance, Target Acquisition, and Reconnaissance (ISTAR) operations. However, the use of UAVs in these applications needs a substantial level of autonomy. In other words, UAVs should have the ability to accomplish planned missions in unexpected situations without requiring human intervention. To ensure this level of autonomy, many artificial intelligence algorithms were designed. These algorithms targeted the guidance, navigation, and control (GNC) of UAVs. In this paper, we described the state of the art of one subset of these algorithms: the deep reinforcement learning (DRL) techniques. We made a detailed description of them, and we deduced the current limitations in this area. We noted that most of these DRL methods were designed to ensure stable and smooth UAV navigation by training computer-simulated environments. We realized that further research efforts are needed to address the challenges that restrain their deployment in real-life scenarios.
引用
收藏
页数:30
相关论文
共 76 条
  • [1] Altawy R, 2017, ACM TRANS CYBER-PHYS, V1, DOI 10.1145/3001836
  • [2] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [3] Azar Ahmad Taher, 2020, 2020 First International Conference of Smart Systems and Emerging Technologies (SMARTTECH), P243, DOI 10.1109/SMART-TECH49988.2020.00061
  • [4] Azar Ahmad Taher, 2021, Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2020. Advances in Intelligent Systems and Computing (AISC 1261), P839, DOI 10.1007/978-3-030-58669-0_74
  • [5] Azar Ahmad Taher, 2021, Proceedings of the International Conference on Advanced Intelligent Systems and Informatics 2020. Advances in Intelligent Systems and Computing (AISC 1261), P862, DOI 10.1007/978-3-030-58669-0_76
  • [6] Bamburry Dane, 2022, Design Management Review, P34, DOI 10.1111/drev.12315
  • [7] UAV Path Planning for Wireless Data Harvesting: A Deep Reinforcement Learning Approach
    Bayerlein, Harald
    Theile, Mirco
    Caccamo, Marco
    Gesbert, David
    [J]. 2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [8] Bohn E, 2019, INT CONF UNMAN AIRCR, P523, DOI [10.1109/icuas.2019.8798254, 10.1109/ICUAS.2019.8798254]
  • [9] Controller Design for Quadrotor UAVs using Reinforcement Learning
    Bou-Ammar, Haitham
    Voos, Holger
    Ertel, Wolfgang
    [J]. 2010 IEEE INTERNATIONAL CONFERENCE ON CONTROL APPLICATIONS, 2010, : 2130 - 2135
  • [10] Bouhamed O., 2020, IEEE INT SYMP CIRC S, P1, DOI [DOI 10.1109/iscas45731.2020.9181245, 10.1109/iscas45731.2020.9181245]