Deep Reinforcement Learning-Based Multichannel Access for Industrial Wireless Networks With Dynamic Multiuser Priority

被引:12
作者
Liu, Xiaoyu [1 ,2 ,3 ,4 ]
Xu, Chi [1 ,2 ,3 ]
Yu, Haibin [1 ,2 ,3 ]
Zeng, Peng [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[2] Chinese Acad Sci, Key Lab Networked Control Syst, Shenyang 110016, Peoples R China
[3] Chinese Acad Sci, Inst Robot & Intelligent Mfg, Shenyang 110169, Peoples R China
[4] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Deep reinforcement learning; dynamic priority; industrial wireless networks (IWNs); multichannel access; quality of service; SPECTRUM ACCESS; COMMUNICATION; TECHNOLOGY; ALLOCATION;
D O I
10.1109/TII.2021.3139349
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In Industry 4.0, massive heterogeneous industrial devices generate a great deal of data with different quality of service requirements, and communicate via industrial wireless networks (IWNs). However, the limited time-frequency resources of IWNs cannot well support the high concurrent access of massive industrial devices with strict real-time and reliable communication requirements. To address this problem, a deep reinforcement learning-based dynamic priority multichannel access (DRL-DPMCA) algorithm is proposed in this article. Firstly, according to the time-sensitivity of industrial data, industrial devices are assigned with different priorities, based on which their channel access probabilities are dynamically adjusted. Then, the Markov decision process is utilized to model the dynamic priority multichannel access problem. To cope with the explosion of state space caused by the multichannel access of massive industrial devices with dynamic priorities, DRL is used to establish the mapping from states to actions. Next, the long-term cumulative reward is maximized to obtain an effective policy. Especially, with joint consideration of the access reward and priority reward, a compound reward for multichannel access and dynamic priority is designed. For breaking the time correlation of training data while accelerating the convergence of DRL-DPMCA, an experience replay with experience-weight is proposed to store and sample experiences categorically. Besides, the gated recurrent unit, dueling architecture and step-by-step epsilon-greedy method are employed to make states more comprehensive and reduce model oscillation. Extensive experiments show that, compared with slotted-Aloha and deep Q network algorithms, DRL-DPMCA converges quickly, and guarantees the highest channel access probability and the minimum queuing delay for high-priority industrial devices in the context of minimum access conflict and nearly 100% channel utilization.
引用
收藏
页码:7048 / 7058
页数:11
相关论文
共 50 条
  • [41] Deep Reinforcement Learning-Based Mode Selection and Resource Management for Green Fog Radio Access Networks
    Sun, Yaohua
    Peng, Mugen
    Mao, Shiwen
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (02): : 1960 - 1971
  • [42] Deep reinforcement learning agents for dynamic spectrum access in television whitespace cognitive radio networks
    Ukpong, Udeme C.
    Idowu-Bismark, Olabode
    Adetiba, Emmanuel
    Kala, Jules R.
    Owolabi, Emmanuel
    Oshin, Oluwadamilola
    Abayomi, Abdultaofeek
    Dare, Oluwatobi E.
    [J]. SCIENTIFIC AFRICAN, 2025, 27
  • [43] Deep Reinforcement Learning-Based Coordinated Beamforming for mmWave Massive MIMO Vehicular Networks
    Tarafder, Pulok
    Choi, Wooyeol
    [J]. SENSORS, 2023, 23 (05)
  • [44] Deep Reinforcement Learning-Based Multi-Access in Massive Machine-Type Communication
    Ravi, Nasim
    Lourenco, Nuno
    Curado, Marilia
    Monteiro, Edmundo
    [J]. IEEE ACCESS, 2024, 12 : 178690 - 178704
  • [45] Deep Reinforcement Learning With Bidirectional Recurrent Neural Networks for Dynamic Spectrum Access
    Chen, Peng
    Quo, Shizeng
    Gao, Yulong
    [J]. 2021 IEEE 94TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2021-FALL), 2021,
  • [46] Deep Reinforcement Learning-Based Routing for Space-Terrestrial Networks
    Tsai, Kai-Chu
    Yao, Ting-Jui
    Huang, Pin-Hao
    Huang, Cheng-Sen
    Han, Zhu
    Wang, Li-Chun
    [J]. 2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [47] Deep Reinforcement Learning-Based Task Scheduling in Heterogeneous MEC Networks
    Shang, Ying
    Li, Jinglei
    Qin, Meng
    Yang, Qinghai
    [J]. 2022 IEEE 95TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-SPRING), 2022,
  • [48] Deep Reinforcement Learning-Based Multipath Routing for LEO Megaconstellation Networks
    Han, Chi
    Xiong, Wei
    Yu, Ronghuan
    [J]. ELECTRONICS, 2024, 13 (15)
  • [49] Deep Reinforcement Learning-Based Intelligent Reflecting Surface for Secure Wireless Communications
    Yang, Helin
    Xiong, Zehui
    Zhao, Jun
    Niyato, Dusit
    Xiao, Liang
    Wu, Qingqing
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (01) : 375 - 388
  • [50] Deep Reinforcement Learning-Based Enhancement of SATMAC for Reliable Channel Access in VANETs
    Wu, Jingbang
    Yu, Ye
    Guo, Yihan
    Zhou, Shufen
    [J]. 2022 IEEE 10TH INTERNATIONAL CONFERENCE ON INFORMATION, COMMUNICATION AND NETWORKS (ICICN 2022), 2022, : 109 - 113