Deep Reinforcement Learning-Based Joint Routing and Capacity Optimization in an Aerial and Terrestrial Hybrid Wireless Network

被引:2
作者
Wang, Zhe [1 ]
Li, Hongxiang [1 ]
Knoblock, Eric J. [2 ]
Apaza, Rafael D. [1 ,2 ]
机构
[1] Univ Louisville, Dept Elect & Comp Engn, Louisville, KY 40292 USA
[2] NASA Glenn Res Ctr, Cleveland, OH 44135 USA
关键词
Delays; Routing; Optimization; Relays; Routing protocols; Vectors; Uplink; ATHN; packet routing; E2E delay; capacity; DRL-based algorithm; D3QN; HOC; INTERNET;
D O I
10.1109/ACCESS.2024.3430191
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As the airspace is experiencing an increasing number of low-altitude aircraft, the concept of spectrum sharing between aerial and terrestrial users emerges as a compelling solution to improve the spectrum utilization efficiency. In this paper, we consider a new Aerial and Terrestrial Hybrid Network (ATHN) comprising aerial vehicles (AVs), ground base stations (BSs), and terrestrial users (TUs). In this ATHN, AVs and BSs collaboratively form a multi-hop ad-hoc network with the objective of minimizing the average end-to-end (E2E) packet transmission delay. Meanwhile, the BSs and TUs form a terrestrial network aimed at maximizing the uplink and downlink sum capacity. Given the concept of spectrum sharing between aerial and terrestrial users in ATHN, we formulate a joint routing and capacity optimization (JRCO) problem, which is a multi-stage combinatorial problem subject to the curse of dimensionality. To address this problem, we propose a Deep Reinforcement Learning (DRL) based algorithm. Specifically, the Dueling Double Deep Q-Network (D3QN) structure is constructed to learn an optimal policy through trial and error. Extensive simulation results demonstrate the efficacy of our proposed solution.
引用
收藏
页码:132056 / 132069
页数:14
相关论文
共 36 条
[1]   Intelligent Resource Allocation in Backscatter-NOMA Networks: A Soft Actor Critic Framework [J].
Alajmi, Abdullah ;
Ahsan, Waleed ;
Fayaz, Muhammad ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (08) :10119-10132
[2]   A New Spectrum Management Concept for Future NAS Communications [J].
Apaza, Rafael D. ;
Knoblock, Eric J. ;
Li, Hongxiang .
2020 AIAA/IEEE 39TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC) PROCEEDINGS, 2020,
[3]   A Deep Reinforcement Learning-Based Geographic Packet Routing Optimization [J].
Bai, Yijie ;
Zhang, Xia ;
Yu, Daojie ;
Li, Shengxiang ;
Wang, Yu ;
Lei, Shuntian ;
Tian, Zhoutai .
IEEE ACCESS, 2022, 10 :108785-108796
[4]   A traffic-aware Q-network enhanced routing protocol based on GPSR for unmanned aerial vehicle ad-hoc networks [J].
Chen, Yi-ning ;
Lyu, Ni-qi ;
Song, Guang-hua ;
Yang, Bo-wei ;
Jiang, Xiao-hong .
FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2020, 21 (09) :1308-1320
[5]   Topology-Aware Resilient Routing Protocol for FANETs: An Adaptive Q-Learning Approach [J].
Cui, Yanpeng ;
Zhang, Qixun ;
Feng, Zhiyong ;
Wei, Zhiqing ;
Shi, Ce ;
Yang, Heng .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (19) :18632-18649
[6]   Multi-agent deep reinforcement learning: a survey [J].
Gronauer, Sven ;
Diepold, Klaus .
ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) :895-943
[7]  
Haarnoja T, 2019, Arxiv, DOI arXiv:1812.05905
[8]  
Hackenberg D. L., 2019, NATL ACAD UAM STUDY
[9]   Joint Velocity and Spectrum Optimization in Urban Air Transportation System via Multi-Agent Deep Reinforcement Learning [J].
Han, Ruixuan ;
Li, Hongxiang ;
Knoblock, Eric J. ;
Gasper, Michael R. ;
Apaza, Rafael D. .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (08) :9770-9782
[10]   Dynamic Spectrum Sharing in Cellular Based Urban Air Mobility via Deep Reinforcement Learning [J].
Han, Ruixuan ;
Li, Hongxiang ;
Knoblock, Eric J. ;
Gasper, Michael R. ;
Apaza, Rafael D. .
2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, :1332-1337