Toward Massive Distribution of Intelligence for 6G Network Management Using Double Deep Q-Networks

被引:1
作者
Majumdar, Sayantini [1 ,2 ]
Schwarzmann, Susanna [1 ]
Trivisonno, Riccardo [1 ]
Carle, Georg [2 ]
机构
[1] Huawei Technol, Munich Res Ctr, D-80992 Munich, Germany
[2] Tech Univ Munich, Dept Informat, D-85748 Garching, Germany
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2024年 / 21卷 / 02期
关键词
6G; network management; network automation; reinforcement learning; machine learning; distributed intelligence; model training stability; scalability; PLACEMENT;
D O I
10.1109/TNSM.2023.3333875
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In future 6G networks, the deployment of network elements is expected to be highly distributed, going beyond the level of distribution of existing 5G deployments. To fully exploit the benefits of such a distributed architecture, there needs to be a paradigm shift from centralized to distributed management. To enable distributed management, Reinforcement Learning (RL) is a promising choice, due to its ability to learn dynamic changes in environments and to deal with complex problems. However, the deployment of highly distributed RL - termed massive distribution of intelligence - still faces a few unsolved challenges. Existing RL solutions, based on Q-Learning (QL) and Deep Q-Network (DQN) do not scale with the number of agents. Therefore, current limitations, i.e., convergence, system performance and training stability, need to be addressed, to facilitate a practical deployment of massive distribution. To this end, we propose improved Double Deep Q-Network (IDDQN), addressing the long-term stability of the agents' training behavior. We evaluate the effectiveness of IDDQN for a beyond 5G/6G use case: auto-scaling virtual resources in a network slice. Simulation results show that IDDQN improves the training stability over DQN and converges at least 2 times sooner than QL. In terms of the number of users served by a slice, IDDQN shows good performance and only deviates on average 8% from the optimal solution. Further, IDDQN is robust and resource-efficient after convergence. We argue that IDDQN is a better alternative than QL and DQN, and holds immense potential for efficiently managing 6G networks.
引用
收藏
页码:2077 / 2094
页数:18
相关论文
共 53 条
  • [1] [Anonymous], 2017, Rep. ETSI GR NFV-IFA 023
  • [2] [Anonymous], 2016, Open source MANO (OSM) information model
  • [3] [Anonymous], 2022, Rep. TS 28.500
  • [4] [Anonymous], 2017, Rep. TR 23.501
  • [5] [Anonymous], 2022, white paper
  • [6] [Anonymous], 2019, Rep. TR 28.533
  • [7] Brown G., 2017, White Paper, V1
  • [8] Zero-Touch AI-Driven Distributed Management for Energy-Efficient 6G Massive Network Slicing
    Chergui, Hatim
    Blanco, Luis
    Garrido, Luis A.
    Ramantas, Kostas
    Kuklinski, Slawomir
    Ksentini, Adlen
    Verikoukis, Christos
    [J]. IEEE NETWORK, 2021, 35 (06): : 43 - 49
  • [9] Cooper A. Feder, 2022, FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, P864, DOI 10.1145/3531146.3533150
  • [10] Cziva R, 2018, IEEE INFOCOM SER, P693, DOI 10.1109/INFOCOM.2018.8486021