Network-Wide Energy-Efficiency Maximization in UAV-Aided IoT Networks: Quasi-Distributed Deep Reinforcement Learning Approach

被引:0
作者
Lee, Seungmin [1 ]
Ban, Tae-Won [2 ]
Lee, Howon [3 ]
机构
[1] Newratek Inc, Dept Wi Fi MAC Stand & Software Engn, Seoul 06175, South Korea
[2] Gyeongsang Natl Univ, Dept AI & Informat Engn, Jinju 52828, South Korea
[3] Ajou Univ, Dept Elect & Comp Engn, Suwon 16499, Gyeonggi, South Korea
基金
新加坡国家研究基金会;
关键词
Q-learning; Internet of Things; Heuristic algorithms; Neural networks; Autonomous aerial vehicles; Three-dimensional displays; Optimization; Batteries; Wireless networks; Trajectory; Multiagent deep reinforcement learning (DRL); network-wide energy efficiency maximization; UAV-base station (BS); UAV Control; uncrewed aerial vehicle (UAV)-aided Internet of Things (IoT) network; DEPLOYMENT;
D O I
10.1109/JIOT.2025.3532477
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In uncrewed aerial vehicle (UAV)-aided Internet of Things (IoT) networks, providing seamless and reliable wireless connectivity to ground devices (GDs) is difficult owing to the short battery lifetimes of UAVs. Hence, we consider a deep reinforcement learning (DRL)-based UAV base station (UAV-BS) control method to maximize the network-wide energy efficiency of UAV-aided IoT networks featuring continuously moving GDs. First, we introduce two centralized DRL approaches; round-robin deep Q-learning (RR-DQL) and selective-k deep Q-learning (SK-DQL), where all UAV-BSs are controlled by a ground control station that collects the status information of UAV-BSs and determines their actions. However, significant signaling overhead and undesired processing latency can occur in these centralized approaches. Hence, we herein propose a quasi-distributed DQL-based UAV-BS control (QD-DQL) method that determines the actions of each agent based on its local information. By performing intensive simulations, we verify the algorithmic robustness and performance excellence of the proposed QD-DQL method based on comparison with several benchmark methods (i.e., RR-DQL, SK-DQL, multiagent Q-learning, and exhaustive search method) while considering the mobility of GDs and the increase in the number of UAV-BSs.
引用
收藏
页码:15404 / 15414
页数:11
相关论文
共 22 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]  
[Anonymous], 2020, Samsung
[3]  
[Anonymous], 2015, ITU-Rec. M.2083-0
[4]  
[Anonymous], 2021, 5G FORUM
[5]   Distributed UAV Deployment in Hostile Environment: A Game-Theoretic Approach [J].
Han, Chen ;
Liu, Aijun ;
An, Kang ;
Zheng, Gan ;
Tong, Xinhai .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2022, 11 (01) :126-130
[6]   HiMAQ: Hierarchical multi-agent Q-learning-based throughput and fairness improvement for UAV-Aided IoT networks [J].
Kim, Eunjin ;
Kim, Junsu ;
Kim, Jae-Hyun ;
Lee, Howon .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2024, 223
[7]   UAV Deployment and IoT Device Association for Energy-Efficient Data-Gathering in Fixed-Wing Multi-UAV Networks [J].
Kuo, Yung-Ching ;
Chiu, Jen-Hao ;
Sheu, Jang-Ping ;
Hong, Y. -W. Peter .
IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2021, 5 (04) :1934-1946
[8]  
Latva-Aho M., 2019, KEY DRIVERS RES CHAL
[9]   Towards 6G Hyper-Connectivity: Vision, Challenges, and Key Enabling Technologies [J].
Lee, Howon ;
Lee, Byungju ;
Yang, Heecheol ;
Kim, Junghyun ;
Kim, Seungnyun ;
Shin, Wonjae ;
Shim, Byonghyo ;
Poor, H. Vincent .
JOURNAL OF COMMUNICATIONS AND NETWORKS, 2023, 25 (03) :344-354
[10]   Optimal Frequency Reuse and Power Control in Multi-UAV Wireless Networks: Hierarchical Multi-Agent Reinforcement Learning Perspective [J].
Lee, Seungmin ;
Lim, Suhyeon ;
Chae, Seong Ho ;
Jung, Bang Chul ;
Park, Chan Yi ;
Lee, Howon .
IEEE ACCESS, 2022, 10 :39555-39565