Multiagent Q-Learning-Based Multi-UAV Wireless Networks for Maximizing Energy Efficiency: Deployment and Power Control Strategy Design

被引:54
作者
Lee, Seungmin [1 ,2 ]
Yu, Heejung [3 ]
Lee, Howon [1 ,2 ]
机构
[1] Hankyong Natl Univ, Sch Elect & Elect Engn, Anseong 17579, South Korea
[2] Hankyong Natl Univ, IITC, Anseong 17579, South Korea
[3] Korea Univ, Dept Elect & Informat Engn, Sejong 30019, South Korea
基金
新加坡国家研究基金会;
关键词
Air-to-ground (A2G) channel; energy efficiency maximization; multiagent distributed Q-learning; power control; unmanned aerial vehicle-base station (UAV-BS);
D O I
10.1109/JIOT.2021.3113128
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In air-to-ground communications, the network lifetime depends on the operation time of unmanned aerial vehicle-base stations (UAV-BSs) owing to the restricted battery capacity. Therefore, the maximization of energy efficiency and the minimization of outage ground users are important metrics of network performance. To achieve these two objectives, the location and transmit power of the UAV-BSs in the network must be optimized This optimization problem may not be tractable in the conventional optimization framework because multiple UAV-BSs interact in a complicated manner. Hence, we formulate the problem as a Markov decision process and develop an algorithm to obtain a solution in a reinforcement learning framework. To avoid a central controller and high computational complexity, we employ a multiagent distributed Q-learning algorithm to obtain a solution. Specifically, we propose a multiagent Q-learning-based UAV-BS deployment and power control strategy to maximize energy efficiency and minimize the number of outage users in multi-UAV wireless networks. Through intensive simulations, it is demonstrated that the proposed algorithm can outperform benchmark algorithms in terms of average energy efficiency and number of average outage users in multi-UAV wireless networks.
引用
收藏
页码:6434 / 6442
页数:9
相关论文
共 18 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]  
[Anonymous], 2019, Key Drivers and Research Challenges for 6G Ubiquitous Wireless Intelligence
[3]  
[Anonymous], 2015, ITU-Rec. M.2083-0
[4]  
[Anonymous], 2012, P14105 ITU
[5]  
Feng QX, 2006, IEEE VTS VEH TECHNOL, P2901
[6]   Energy-Efficient UAV-Enabled Data Collection via Wireless Charging: A Reinforcement Learning Approach [J].
Fu, Shu ;
Tang, Yujie ;
Wu, Yuan ;
Zhang, Ning ;
Gu, Huaxi ;
Chen, Chen ;
Liu, Min .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (12) :10209-10219
[7]   6G: Opening New Horizons for Integration of Comfort, Security, and Intelligence [J].
Gui, Guan ;
Liu, Miao ;
Tang, Fengxiao ;
Kato, Nei ;
Adachi, Fumiyuki .
IEEE WIRELESS COMMUNICATIONS, 2020, 27 (05) :126-132
[8]   Distributed Drone Base Station Positioning for Emergency Cellular Networks Using Reinforcement Learning [J].
Klaine, Paulo V. ;
Nadas, Joao P. B. ;
Souza, Richard D. ;
Imran, Muhammad A. .
COGNITIVE COMPUTATION, 2018, 10 (05) :790-804
[9]  
Lauer M., 2000, P 17 INT C MACHINE L, P535
[10]  
Lee S., 2020, P KICS WINT C FEB 20, P870