Toward Energy-Efficient Dynamic Resource Allocation in Uplink NOMA Systems: Deep Reinforcement Learning for Single and Multi-Cell NOMA Systems

被引:0
作者
Rabee, Ayman [1 ]
Barhumi, Imad [1 ]
机构
[1] United Arab Emirates Univ, EECE Dept, Al Ain 15551, U Arab Emirates
关键词
Resource management; NOMA; Optimization; Uplink; Heuristic algorithms; Wireless networks; Wireless communication; Quality of service; Dynamic scheduling; Vehicle dynamics; Nonorthogonal multiple access (NOMA); deep reinforcement learning; deep Q-network; twin delayed deep deterministic policy gradient; NONORTHOGONAL MULTIPLE-ACCESS; POWER ALLOCATION; MASSIVE MIMO; NETWORKS; OPTIMIZATION;
D O I
10.1109/TVT.2025.3540940
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Non-orthogonal multiple access (NOMA) is a key technology for future wireless networks, enabling improved spectral efficiency and massive device connectivity. However, optimizing power allocation, subchannel assignment, and cell association in multi-cell uplink NOMA systems is challenging due to user mobility and the NP-hard nature of the problem. This paper addresses these challenges by formulating the problem as a mixed-integer non-linear programming (MINLP) model to maximize energy efficiency (EE). We propose a deep reinforcement learning framework that employs deep Q-networks (DQN) for cell association and subchannel assignment, and twin delayed deep deterministic policy gradient (TD3) for power allocation. Simulation results reveal significant EE improvements, with multi-agent TD3 (MATD3) outperforming traditional Lagrange methods and multi-agent deep deterministic policy gradient (MADDPG). Furthermore, the proposed method exhibits robust adaptability to user mobility and superior performance in multi-cell environments, effectively mitigating inter-cell interference and enhancing resource allocation in dynamic scenarios.
引用
收藏
页码:9313 / 9327
页数:15
相关论文
共 51 条
[1]  
3rd Generation Partnership Project (3GPP), 2014, Tech. Specification 36.931, V5
[2]   Resource Allocation in Uplink NOMA-IoT Networks: A Reinforcement-Learning Approach [J].
Ahsan, Waleed ;
Yi, Wenqiang ;
Qin, Zhijin ;
Liu, Yuanwei ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (08) :5083-5098
[3]   Joint user association and resource allocation for cost-efficient NOMA-enabled F-RANs [J].
Ai, Yuan ;
Liu, Chenxi ;
Peng, Mugen .
DIGITAL COMMUNICATIONS AND NETWORKS, 2024, 10 (06) :1686-1697
[4]   GEKKO Optimization Suite [J].
Beal, Logan D. R. ;
Hill, Daniel C. ;
Martin, R. Abraham ;
Hedengren, John D. .
PROCESSES, 2018, 6 (08)
[5]  
Bi ZG, 2020, IEEE VTS VEH TECHNOL
[6]  
Boyd S., 2004, Convex optimization, DOI 10.1017/CBO9780511804441
[7]   On the Maximum Energy Efficiency of Random Access-Based OMA and NOMA in Multirate Environment [J].
Cao, Shengbin ;
Hou, Fen .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (12) :10438-10454
[8]   Deep Reinforcement Learning for Network Selection Over Heterogeneous Health Systems [J].
Chkirbene, Zina ;
Abdellatif, Alaa Awad ;
Mohamed, Amr ;
Erbad, Aiman ;
Guizani, Mohsen .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2022, 9 (01) :258-270
[9]   Energy-Efficient Resource Allocation for Heterogeneous SWIPT-NOMA Systems [J].
Cui, Haixia ;
Ye, Xianwan ;
You, Fan .
IEEE ACCESS, 2022, 10 :79281-79288
[10]   A Survey of Non-Orthogonal Multiple Access for 5G [J].
Dai, Linglong ;
Wang, Bichai ;
Ding, Zhiguo ;
Wang, Zhaocheng ;
Chen, Sheng ;
Hanzo, Lajos .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2018, 20 (03) :2294-2323