Deep Reinforcement Learning-Based Multichannel Access for Industrial Wireless Networks With Dynamic Multiuser Priority

被引:12
作者
Liu, Xiaoyu [1 ,2 ,3 ,4 ]
Xu, Chi [1 ,2 ,3 ]
Yu, Haibin [1 ,2 ,3 ]
Zeng, Peng [1 ,2 ,3 ]
机构
[1] Chinese Acad Sci, Shenyang Inst Automat, State Key Lab Robot, Shenyang 110016, Peoples R China
[2] Chinese Acad Sci, Key Lab Networked Control Syst, Shenyang 110016, Peoples R China
[3] Chinese Acad Sci, Inst Robot & Intelligent Mfg, Shenyang 110169, Peoples R China
[4] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
Deep reinforcement learning; dynamic priority; industrial wireless networks (IWNs); multichannel access; quality of service; SPECTRUM ACCESS; COMMUNICATION; TECHNOLOGY; ALLOCATION;
D O I
10.1109/TII.2021.3139349
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In Industry 4.0, massive heterogeneous industrial devices generate a great deal of data with different quality of service requirements, and communicate via industrial wireless networks (IWNs). However, the limited time-frequency resources of IWNs cannot well support the high concurrent access of massive industrial devices with strict real-time and reliable communication requirements. To address this problem, a deep reinforcement learning-based dynamic priority multichannel access (DRL-DPMCA) algorithm is proposed in this article. Firstly, according to the time-sensitivity of industrial data, industrial devices are assigned with different priorities, based on which their channel access probabilities are dynamically adjusted. Then, the Markov decision process is utilized to model the dynamic priority multichannel access problem. To cope with the explosion of state space caused by the multichannel access of massive industrial devices with dynamic priorities, DRL is used to establish the mapping from states to actions. Next, the long-term cumulative reward is maximized to obtain an effective policy. Especially, with joint consideration of the access reward and priority reward, a compound reward for multichannel access and dynamic priority is designed. For breaking the time correlation of training data while accelerating the convergence of DRL-DPMCA, an experience replay with experience-weight is proposed to store and sample experiences categorically. Besides, the gated recurrent unit, dueling architecture and step-by-step epsilon-greedy method are employed to make states more comprehensive and reduce model oscillation. Extensive experiments show that, compared with slotted-Aloha and deep Q network algorithms, DRL-DPMCA converges quickly, and guarantees the highest channel access probability and the minimum queuing delay for high-priority industrial devices in the context of minimum access conflict and nearly 100% channel utilization.
引用
收藏
页码:7048 / 7058
页数:11
相关论文
共 50 条
  • [31] A Dueling Deep Recurrent Q-Network Framework for Dynamic Multichannel Access in Heterogeneous Wireless Networks
    Chen, Haitao
    Zhao, Haitao
    Zhou, Li
    Zhang, Jiao
    Liu, Yan
    Pan, Xiaoqian
    Liu, Xingguang
    Wei, Jibo
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2022, 2022
  • [32] Federated Deep Reinforcement Learning-Based Spectrum Access Algorithm With Warranty Contract in Intelligent Transportation Systems
    Zhu, Rongbo
    Li, Mengyao
    Liu, Hao
    Liu, Lu
    Ma, Maode
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (01) : 1178 - 1190
  • [33] Deep Reinforcement Learning-based Scheduling for Roadside Communication Networks
    Atallah, Rihal
    Assi, Chadi
    Khahhaz, Maurice
    2017 15TH INTERNATIONAL SYMPOSIUM ON MODELING AND OPTIMIZATION IN MOBILE, AD HOC, AND WIRELESS NETWORKS (WIOPT), 2017,
  • [34] Resilient Dynamic Channel Access via Robust Deep Reinforcement Learning
    Wang, Feng
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    IEEE ACCESS, 2021, 9 : 163188 - 163203
  • [35] Deep Reinforcement Learning-Based Adaptive Scheduling for Wireless Time-Sensitive Networking
    Kim, Hanjin
    Kim, Young-Jin
    Kim, Won-Tae
    SENSORS, 2024, 24 (16)
  • [36] Deep Reinforcement Learning-Based High Concurrent Computing Offloading for Heterogeneous Industrial Tasks
    Liu X.-Y.
    Xu C.
    Zeng P.
    Yu H.-B.
    Jisuanji Xuebao/Chinese Journal of Computers, 2021, 44 (12): : 2367 - 2381
  • [37] A deep reinforcement learning-based multi-optimality routing scheme for dynamic IoT networks
    Cong, Peizhuang
    Zhang, Yuchao
    Liu, Zheli
    Baker, Thar
    Tawfik, Hissam
    Wang, Wendong
    Xu, Ke
    Li, Ruidong
    Li, Fuliang
    COMPUTER NETWORKS, 2021, 192
  • [38] GTD3-NET: A deep reinforcement learning-based routing optimization algorithm for wireless networks
    Lu, You
    Zhang, Zhe
    Xu, Xiao
    Liu, Lanhui
    Fu, Qiming
    Chen, Jianping
    Chen, Chen
    PEER-TO-PEER NETWORKING AND APPLICATIONS, 2025, 18 (01) : 16 - 16
  • [39] Deceiving Reactive Jamming in Dynamic Wireless Sensor Networks: A Deep Reinforcement Learning Based Approach
    Zhang, Chen
    Mao, Tianqi
    Xiao, Zhenyu
    Liu, Ruiqi
    Xia, Xiang-Gen
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 4455 - 4460
  • [40] Dynamic Reservation and Deep Reinforcement Learning based Autonomous Resource Management for Wireless Virtual Networks
    Sun, Guolin
    Zemuy, Gebrekidan Tesfay
    Xiong, Kun
    2018 IEEE 37TH INTERNATIONAL PERFORMANCE COMPUTING AND COMMUNICATIONS CONFERENCE (IPCCC), 2018,