Deep Q-Networks for Aerial Data Collection in Multi-UAV-Assisted Wireless Sensor Networks

被引:11
作者
Emami, Yousef [1 ]
Wei, Bo [2 ]
Li, Kai [1 ]
Ni, Wei [3 ]
Tovar, Eduardo [1 ]
机构
[1] CISTER Res Ctr, Porto, Portugal
[2] Northumbria Univ, Newcastle, England
[3] CSIRO, Sydney, NSW, Australia
来源
IWCMC 2021: 2021 17TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE (IWCMC) | 2021年
关键词
Unmanned aerial vehicles; Communication scheduling; Multi-UAV Deep Reinforcement Learning; Deep Q-Network;
D O I
10.1109/IWCMC51323.2021.9498726
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned Aerial Vehicles (UAVs) can collaborate to collect and relay data for ground sensors in remote and hostile areas. In multi-UAV-assisted wireless sensor networks (MA-WSN), the UAVs' movements impact on channel condition and can fail data transmission, this situation along with newly arrived data give rise to buffer overflows at the ground sensors. Thus, scheduling data transmission is of utmost importance in MA-WSN to reduce data packet losses resulting from buffer overflows and channel fading. In this paper, we investigate the optimal ground sensor selection at the UAVs to minimize data packet losses. The optimization problem is formulated as a multi-agent Markov decision process, where network states consist of battery levels and data buffer lengths of the ground sensor, channel conditions, and waypoints of the UAV along the trajectory. In practice, an MA-WSN contains a large number of network states, while the up-to-date knowledge of the network states and other UAVs' sensor selection decisions is not available at each agent. We propose a Multi-UAV Deep Reinforcement Learning based Scheduling Algorithm (MUAIS) to minimize the data packet loss, where the UAVs learn the underlying patterns of the data and energy arrivals at all the ground sensors. Numerical results show that the proposed MUAIS achieves at least 46% and 35% lower packet loss than an optimal solution with single-UAV and an existing non-learning greedy algorithm, respectively.
引用
收藏
页码:669 / 674
页数:6
相关论文
共 20 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]  
Challita U, 2018, IEEE ICC
[3]   Multi-Agent Reinforcement Learning-Based Resource Allocation for UAV Networks [J].
Cui, Jingjing ;
Liu, Yuanwei ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (02) :729-743
[4]  
Emami Y, 2020, 2020 IEEE 91 VEH TEC, P1
[5]  
Gao YT, 2020, PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), P2175, DOI [10.1109/ITNEC48623.2020.9084954, 10.1109/itnec48623.2020.9084954]
[6]   Small unmanned airborne systems to support oil and gas pipeline monitoring and mapping [J].
Gomez, Cristina ;
Green, David R. .
ARABIAN JOURNAL OF GEOSCIENCES, 2017, 10 (09)
[7]   Onboard Deep Deterministic Policy Gradients for Online Flight Resource Allocation of UAVs [J].
Li, Kai ;
Emami, Yousef ;
Ni, Wei ;
Tovar, Eduardo ;
Han, Zhu .
IEEE Networking Letters, 2020, 2 (03) :106-110
[8]   Unmanned Aerial Vehicles in Agriculture: A Review of Perspective of Platform, Control, and Applications [J].
Kim, Jeongeun ;
Kim, Seungwon ;
Ju, Chanyoung ;
Son, Hyoung Il .
IEEE ACCESS, 2019, 7 (105100-105115) :105100-105115
[9]   Online Velocity Control and Data Capture of Drones for the Internet of Things: An Onboard Deep Reinforcement Learning Approach [J].
Li, Kai ;
Ni, Wei ;
Tovard, Eduardo ;
Jamalipour, Abbas .
IEEE VEHICULAR TECHNOLOGY MAGAZINE, 2021, 16 (01) :49-56
[10]   On-Board Deep Q-Network for UAV-Assisted Online Power Transfer and Data Collection [J].
Li, Kai ;
Ni, Wei ;
Tovar, Eduardo ;
Jamalipour, Abbas .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (12) :12215-12226