RETRACTED: Energy-aware resource management for uplink non-orthogonal multiple access: Multi-agent deep reinforcement learning (Retracted Article)

被引:5
作者
Li, Yingfang [1 ]
Yang, Bo [1 ]
Yan, Li [1 ]
Gao, Wei [2 ]
机构
[1] Honghe Univ, Sch Engn, Mengzi 661199, Peoples R China
[2] Yunnan Normal Univ, Sch Informat Sci & Technol, Kunming 650500, Yunnan, Peoples R China
来源
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE | 2020年 / 105卷
关键词
Non-orthogonal multiple access; Resource allocation; Energy efficiency; Deep reinforcement learning; Deep deterministic policy gradient; POWER ALLOCATION; NOMA SYSTEMS; 5G SYSTEMS; OPPORTUNITIES;
D O I
10.1016/j.future.2019.12.047
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Non-orthogonal multiple access (NOMA) is one of the promising technologies to meet the huge access demand and the high data rate requirements of the next generation networks. In this paper, we investigate the joint subchannel assignment and power allocation problem in an uplink multi-user NOMA system to maximize the energy efficiency (EE) while ensuring the quality-of-service (QoS) of all users. Different from conventional model-based resource allocation methods, we propose two deep reinforcement learning (DRL) based frameworks to solve this non-convex and dynamic optimization problem, referred to as discrete DRL based resource allocation (DDRA) framework and continuous DRL based resource allocation (CDRA) framework. Specifically, for the DDRA framework, we use a deep Q network (DQN) to output the optimum subchannel assignment policy, and design a distributed and discretized multi-DQN based network to allocate the corresponding transmit power of all users. For the CDRA framework, we design a joint DQN and deep deterministic policy gradient (DDPG) based network to generate the optimal subchannel assignment and power allocation policy. The entire resource allocation policies of these two frameworks are adjusted by updating the weights of their neural networks according to feedback of the system. Numerical results show that the proposed DRL-based resource allocation frameworks can significantly improve the EE of the whole NOMA system compared with other approaches. The proposed DRL based frameworks can provide good performance in various moving speed scenarios through adjusting learning parameters. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:684 / 694
页数:11
相关论文
共 36 条
  • [21] Human-level control through deep reinforcement learning
    Mnih, Volodymyr
    Kavukcuoglu, Koray
    Silver, David
    Rusu, Andrei A.
    Veness, Joel
    Bellemare, Marc G.
    Graves, Alex
    Riedmiller, Martin
    Fidjeland, Andreas K.
    Ostrovski, Georg
    Petersen, Stig
    Beattie, Charles
    Sadik, Amir
    Antonoglou, Ioannis
    King, Helen
    Kumaran, Dharshan
    Wierstra, Daan
    Legg, Shane
    Hassabis, Demis
    [J]. NATURE, 2015, 518 (7540) : 529 - 533
  • [22] An Introduction to Deep Learning for the Physical Layer
    O'Shea, Timothy
    Hoydis, Jakob
    [J]. IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2017, 3 (04) : 563 - 575
  • [23] Sen S., 2010, P 9 ACM SIGCOMM WORK, p17:1
  • [24] Sutton RS, 2018, ADAPT COMPUT MACH LE, P1
  • [25] Sutton RS, 2000, ADV NEUR IN, V12, P1057
  • [26] Wang CL, 2018, PRECIS AGRIC, V19, P1062, DOI 10.1007/s11119-018-9574-5
  • [27] Deep Learning for Wireless Physical Layer: Opportunities and Challenges
    Wang, Tianqi
    Wen, Chao-Kai
    Wang, Hanqing
    Gao, Feifei
    Jiang, Tao
    Jin, Shi
    [J]. CHINA COMMUNICATIONS, 2017, 14 (11) : 92 - 111
  • [28] Low-Complexity Power Allocation in NOMA Systems With Imperfect SIC for Maximizing Weighted Sum-Rate
    Wang, Xiaoming
    Chen, Ruilu
    Xu, Youyun
    Meng, Qingmin
    [J]. IEEE ACCESS, 2019, 7 : 94238 - 94253
  • [29] WATKINS CJCH, 1992, MACH LEARN, V8, P279, DOI 10.1007/BF00992698
  • [30] Reinforcement Learning-Based NOMA Power Allocation in the Presence of Smart Jamming
    Xiao, Liang
    Li, Yanda
    Dai, Canhuang
    Dai, Huaiyu
    Poor, H. Vincent
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2018, 67 (04) : 3377 - 3389