Radio Resource Management for Cellular-Connected UAV: A Learning Approach

被引:30
作者
Li, Yuanjian [1 ]
Aghvami, A. Hamid [1 ]
机构
[1] Kings Coll London, Ctr Telecommun Res CTR, London WC2R 2LS, England
关键词
Interference; Autonomous aerial vehicles; Resource management; Channel models; Cellular networks; Array signal processing; Buildings; Unmanned aerial vehicle (UAV); cellular networks; deep reinforcement learning; interference management; beamforming; ENERGY-EFFICIENT; INTERFERENCE COORDINATION; COMMUNICATION; ALLOCATION; NOMA; OPTIMIZATION; CANCELLATION; STRESS; CHINA;
D O I
10.1109/TCOMM.2023.3262826
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Integrating unmanned aerial vehicles (UAVs) into existing cellular networks encounters lots of challenges, among which one of the most striking concerns is how to achieve harmonious coexistence of aerial transceivers, inter alia, UAVs, and terrestrial user equipments (UEs). In this paper, a cellular-connected UAV network is focused, where multiple UAVs receive messages from base stations (BSs) in the down-link, while BSs are serving ground UEs in their cells. For effectively managing inter-cell interferences (ICIs) among UEs due to intense reuse of time-frequency resource block (RB) resource, a first $p$ -tier based RB coordination criterion is proposed and adopted. Then, to enhance wireless transmission quality for UAVs while protecting terrestrial UEs from being interfered by ground-to-air (G2A) transmissions, a radio resource management (RRM) problem of joint dynamic RB coordination and time-varying beamforming design minimizing UAV's ergodic outage duration (EOD) is investigated. To cope with conventional optimization techniques' inefficiency in solving the formulated RRM problem, a deep reinforcement learning (DRL)-aided solution is initiated, where deep double duelling Q network (D3QN) and twin delayed deep deterministic policy gradient (TD3) are invoked to deal with RB coordination in discrete action domain and beamforming design in continuous action regime, respectively. The hybrid D3QN-TD3 solution is trained via interacting with the considered outer and inner environments in an online centralized manner so that it can then help achieve the suboptimal EOD minimization performance during its offline decentralized exploitation. Simulation results have illustrated the effectiveness of the proposed hybrid D3QN-TD3 algorithm, compared to several representative baselines.
引用
收藏
页码:2784 / 2800
页数:17
相关论文
共 46 条
[1]   Green internet of things using UAVs in B5G networks: A review of applications and strategies [J].
Alsamhi, S. H. ;
Afghah, Fatemeh ;
Sahal, Radhya ;
Hawbani, Ammar ;
Al-qaness, Mohammed A. A. ;
Lee, B. ;
Guizani, Mohsen .
AD HOC NETWORKS, 2021, 117
[2]   Federated Learning Over Wireless Fading Channels [J].
Amiri, Mohammad Mohammadi ;
Gunduz, Deniz .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (05) :3546-3557
[3]  
[Anonymous], 2017, document TR 36.777, Version 15.0.0
[4]  
[Anonymous], 2012, ITU-Rec. P. 530-14
[5]   Interference Coordination and Cancellation for 4G Networks [J].
Boudreau, Gary ;
Panicker, John ;
Guo, Ning ;
Chang, Rui ;
Wang, Neng ;
Vrzic, Sophie .
IEEE COMMUNICATIONS MAGAZINE, 2009, 47 (04) :74-81
[6]   Massive MIMO for Communications With Drone Swarms [J].
Chandhar, Prabhu ;
Danev, Danyo ;
Larsson, Erik G. .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (03) :1604-1629
[8]   UAV Assisted Spectrum Sharing Ultra-Reliable and Low-Latency Communications [J].
Chu, Zheng ;
Hao, Wanming ;
Xiao, Pei ;
Jia Shia .
2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
[9]   3D UAV Trajectory Design and Frequency Band Allocation for Energy-Efficient and Fair Communication: A Deep Reinforcement Learning Approach [J].
Ding, Ruijin ;
Gao, Feifei ;
Shen, Xuemin .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (12) :7796-7809
[10]   Cell-Free Massive MIMO: A Survey [J].
Elhoushy, Salah ;
Ibrahim, Mohamed ;
Hamouda, Walaa .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2022, 24 (01) :492-523