A Reinforcement Learning Approach for D2D Spectrum Sharing in Wireless Industrial URLLC Networks

被引:0
作者
Sanusi, Idayat O. [1 ]
Nasr, Karim M. [1 ]
机构
[1] Univ Greenwich, Fac Engn & Sci, London ME4 4TB, Kent, England
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2024年 / 21卷 / 05期
关键词
Quality of service; Resource management; Throughput; Wireless communication; Interference; Fifth generation (5G) and beyond wireless networks; radio spectrum management (RRM); distributed algorithms; device-to-device communication (D2D); reinforcement learning; matching theory; TO-DEVICE COMMUNICATIONS; RESOURCE-ALLOCATION; MATCHING THEORY; COMMUNICATION;
D O I
10.1109/TNSM.2024.3445123
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distributed Radio Resource Management (RRM) solutions are gaining an increasing interest recently, especially when a large number of devices are present as in the case of a wireless industrial network. Self-organisation relying on distributed RRM schemes is envisioned to be one of the key pillars of 5G and beyond Ultra Reliable Low Latency Communication (URLLC) networks. Reinforcement learning is emerging as a powerful distributed technique to facilitate self-organisation. In this paper, spectrum sharing in a Device-to-Device (D2D)-enabled wireless network is investigated, targeting URLLC applications. A distributed scheme denoted as Reinforcement Learning Based Matching (RLBM) which combines reinforcement learning and matching theory, is presented with the aim of achieving an autonomous device-based resource allocation. A distributed local Q-table is used to avoid global information gathering and a stateless Q-learning approach is adopted, therefore reducing requirements for a large state-action mapping. Simulation case studies are used to verify the performance of the presented approach in comparison with other RRM techniques. The presented RLBM approach results in a good tradeoff of throughput, complexity and signalling overheads while maintaining the target Quality of Service/Experience (QoS/QoE) requirements of the different users in the network.
引用
收藏
页码:5410 / 5419
页数:10
相关论文
共 36 条
  • [1] 5g-ppp, White paper
  • [2] [Anonymous], 2010, Rep. 36.814
  • [3] [Anonymous], 2015, CISC VIS NETW IND GL
  • [4] An Autonomous Learning-Based Algorithm for Joint Channel and Power Level Selection by D2D Pairs in Heterogeneous Cellular Networks
    Asheralieva, Alia
    Miyanaga, Yoshikazu
    [J]. IEEE TRANSACTIONS ON COMMUNICATIONS, 2016, 64 (09) : 3996 - 4012
  • [5] Matching Theory Applications in wireless communications
    Bayat, Siavash
    Li, Yonghui
    Song, Lingyang
    Han, Zhu
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2016, 33 (06) : 103 - 122
  • [6] A Survey on Machine-Learning Techniques in Cognitive Radios
    Bkassiny, Mario
    Li, Yang
    Jayaweera, Sudharman K.
    [J]. IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2013, 15 (03): : 1136 - 1159
  • [7] De Bast S, 2019, IEEE CONF COMPUT, P264, DOI [10.1109/infcomw.2019.8845211, 10.1109/INFCOMW.2019.8845211]
  • [8] Feki S, 2019, INT WIREL COMMUN, P1367, DOI [10.1109/iwcmc.2019.8766509, 10.1109/IWCMC.2019.8766509]
  • [9] Learning Multirobot Hose Transportation and Deployment by Distributed Round-Robin Q-Learning
    Fernandez-Gauna, Borja
    Etxeberria-Agiriano, Ismael
    Grana, Manuel
    [J]. PLOS ONE, 2015, 10 (07):
  • [10] Deep Multiagent Reinforcement-Learning-Based Resource Allocation for Internet of Controllable Things
    Gu, Bo
    Zhang, Xu
    Lin, Ziqi
    Alazab, Mamoun
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (05) : 3066 - 3074