Multi-Objective Reinforcement Learning for Power Allocation in Massive MIMO Networks: A Solution to Spectral and Energy Trade-Offs

被引:0
作者
Oh, Youngwoo [1 ]
Ullah, Arif [1 ]
Choi, Wooyeol [1 ]
机构
[1] Chosun Univ, Coll IT Convergence, Dept Comp Engn, Gwangju 61452, South Korea
关键词
5G and beyond networks; energy efficiency; massive MIMO; multi-objective reinforcement learning; power allocation; spectral efficiency;
D O I
10.1109/ACCESS.2023.3347788
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The joint optimization of spectral efficiency (SE) and energy efficiency (EE) through power allocation (PA) techniques is a critical requirement for emerging fifth-generation and beyond networks. The trade-off between SE and EE becomes challenging in the massive multiple-input-multiple-output (MIMO) equipped base stations (BSs) in multi-cell cellular networks. Various algorithmic approaches including genetic algorithms and convex optimization have been considered to optimize the trade-offs between SE and EE in cellular networks. However, these methods suffer from high computational costs. A promising deep reinforcement learning technique is capable of addressing the computational challenges of single-objective optimization problems in wireless networks. Furthermore, multi-objective reinforcement learning has been employed for multi-objective optimization problems and can be utilized to jointly enhance the SE and EE in cellular networks. In this paper, we propose a downlink (DL) transmit PA method based on a multi-objective asynchronous advantage single actor-multiple critics (MO-A3Cs) architecture. The proposed architecture aims to optimize SE and EE trade-offs in massive MIMO-assisted multi-cell networks. Furthermore, we also propose a Bayesian rule-based preference weight updating mechanism, multi-objective advantage function, and balanced-reward aggregation method to effectively train and avoid biased objective reward during the training process of the proposed model. Extensive simulations depict that the proposed model is better capable of dealing with the joint optimization of SE and EE in dynamic changing scenarios. Compared to the existing benchmarks such as Pareto front approximation-based multi-objective, reinforcement learning-based single objective, and iterative methods, the proposed approach provides a better SE-EE trade-off by achieving a higher EE in multi-cell massive MIMO networks.
引用
收藏
页码:1172 / 1188
页数:17
相关论文
共 58 条
  • [1] Abels A, 2019, PR MACH LEARN RES, V97
  • [2] An Efficient Pilot Assignment Scheme for Addressing Pilot Contamination in Multicell Massive MIMO Systems
    Al-hubaishi, Ahmed S.
    Noordin, Nor Kamariah
    Sali, Aduwati
    Subramaniam, Shamala
    Mansoor, Ali Mohammed
    [J]. ELECTRONICS, 2019, 8 (04):
  • [3] Energy Efficiency-Spectral Efficiency Tradeoff: A Multiobjective Optimization Approach
    Amin, Osama
    Bedeer, Ebrahim
    Ahmed, Mohamed Hossam
    Dobre, Octavia A.
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2016, 65 (04) : 1975 - 1981
  • [4] Training Effect on AI-based Resource Allocation in small-cell networks
    Anzaldo, Alexis
    Andrade, Angel G.
    [J]. 2021 IEEE LATIN-AMERICAN CONFERENCE ON COMMUNICATIONS (LATINCOM 2021), 2021,
  • [5] Recent Advances in Hierarchical Reinforcement Learning
    Andrew G. Barto
    Sridhar Mahadevan
    [J]. Discrete Event Dynamic Systems, 2003, 13 (4) : 341 - 379
  • [6] Basaklar T., 2022, PROC 11 INT C LEARN
  • [7] Massive MIMO Has Unlimited Capacity
    Bjornson, Emil
    Hoydis, Jakob
    Sanguinetti, Luca
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (01) : 574 - 590
  • [8] Björnson E, 2017, FOUND TRENDS SIGNAL, V11, P154, DOI 10.1561/2000000093
  • [9] Chen RQ, 2021, EUR SIGNAL PR CONF, P1631, DOI 10.23919/EUSIPCO54536.2021.9615934
  • [10] Multi-Objective Power Minimization Design for Energy Efficiency in Multicell Multiuser MIMO Beamforming System
    Chen, Wei-Yu
    Hsieh, Po-Ya
    Chen, Bor-Sen
    [J]. IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2020, 4 (01): : 31 - 45