Reinforcement Learning for Joint Detection and Mapping Using Dynamic UAV Networks

被引:5
作者
Guerra, Anna [1 ]
Guidi, Francesco [1 ]
Dardari, Davide [2 ]
Djuric, Petar M. [3 ]
机构
[1] Natl Res Council Italy, CNR IEIIT, I-40136 Bologna, Italy
[2] Univ Bologna, WiLAB DEI Guglielmo Marconi CNIT, I-40136 Bologna, Italy
[3] SUNY Stony Brook, ECE, Stony Brook, NY 11794 USA
关键词
Autonomous aerial vehicles; Sensors; Task analysis; Navigation; Q-learning; Radar; State estimation; Autonomous navigation; reinforcement learning (RL); task assignment; unmanned aerial vehicles (UAVs); SWARM; LOCALIZATION; NAVIGATION; DESIGN; SENSOR; 5G;
D O I
10.1109/TAES.2023.3300813
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
Dynamic radar networks (DRNs), usually composed of flying unmanned aerial vehicles (UAVs), have recently attracted great interest for time-critical applications, such as search-and-rescue operations, involving reliable detection of multiple targets and situational awareness through environment radio mapping. Unfortunately, the time available for detection is often limited, and in most settings, there are no reliable models of the environment, which should be learned quickly. One possibility to guarantee short learning time is to enhance cooperation among UAVs. For example, they can share information for properly navigating the environment if they have a common goal. Alternatively, in case of multiple and different goals or tasks, they can exchange their available information to fitly assign tasks (e.g., targets) to each network agent. In this article, we consider ad hoc approaches for task assignment and a multi-agent reinforcement learning algorithm that allow the UAVs to learn a suitable navigation policy to explore an unknown environment while maximizing the accuracy in detecting targets. The obtained results demonstrate that cooperation at different levels accelerates the learning process and brings benefits in accomplishing the team's goals.
引用
收藏
页码:2586 / 2601
页数:16
相关论文
共 82 条
  • [1] Abed-Alguni Bilal H., 2016, International Journal of Artificial Intelligence, V14, P71
  • [2] Expertness based cooperative Q-learning
    Ahmadabadi, MN
    Asadpour, M
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2002, 32 (01): : 66 - 76
  • [3] A Reinforcement Learning Based Approach for Multitarget Detection in Massive MIMO Radar
    Ahmed, Aya Mostafa
    Ahmad, Alaa Alameer
    Fortunati, Stefano
    Sezgin, Aydin
    Greco, Maria Sabrina
    Gini, Fulvio
    [J]. IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2021, 57 (05) : 2622 - 2636
  • [4] Annasamy RM, 2019, AAAI CONF ARTIF INTE, P4561
  • [5] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [6] Multi-UAV Path Planning for Wireless Data Harvesting With Deep Reinforcement Learning
    Bayerlein, Harald
    Theile, Mirco
    Caccamo, Marco
    Gesbert, David
    [J]. IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2021, 2 : 1171 - 1187
  • [7] Bayerlein H, 2018, IEEE INT WORK SIGN P, P945
  • [8] Bertsekas DP, 1995, PROCEEDINGS OF THE 34TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-4, P560, DOI 10.1109/CDC.1995.478953
  • [9] Merging occupancy grid maps from multiple robots
    Birk, Andreas
    Carpin, Stefano
    [J]. PROCEEDINGS OF THE IEEE, 2006, 94 (07) : 1384 - 1397
  • [10] Brys T, 2014, IEEE IJCNN, P2315, DOI 10.1109/IJCNN.2014.6889732