Multiagent Reinforcement Learning Based Distributed Channel Access for Industrial Edge-Cloud Web 3.0

被引:3
作者
Yang, Chen [1 ]
Wang, Yushi [2 ]
Lan, Shulin [3 ]
Zhu, Liehuang [1 ]
机构
[1] Beijing Inst Technol, Sch Cyberspace Sci & Technol, Beijing 100081, Peoples R China
[2] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing 100081, Peoples R China
[3] Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100081, Peoples R China
来源
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING | 2024年 / 11卷 / 05期
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Wireless communication; Wireless sensor networks; Smart manufacturing; Task analysis; Reinforcement learning; Resource management; Production facilities; Smart factory; channel access; multiagent reinforcement learning; edge-cloud collaboration; edge computing; ALLOCATION; NETWORKS;
D O I
10.1109/TNSE.2024.3377441
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In the emerging Web 3.0 applications for mass customized and personalized manufacturing, smart mobile resources need to interact with each other and other resources to achieve efficient collaborative manufacturing. Existing wireless communication solutions cannot leverage multiantenna technology and the movement direction of smart mobile resources to meet the high requirements for communication rate and reliability in high-performance manufacturing processes. Therefore, this paper proposes a task-aware distributed channel access scheme for multiantenna smart mobile resources in a factory. First, this paper introduces an edge-cloud collaboration framework for smart factories to support autonomous wireless access point selection for mobile resources. Second, a user-centric active wireless channel access scheme is proposed and a channel resource allocation optimization problem is formulated for mobile resources to leverage multiple antennas and movement direction to address the unstable connection problem. Third, a centralized-training-and-distributed-execution multiagent reinforcement learning (MARL) model with a specially designed neural network architecture is built for smart mobile resources, effectively using important input information of the next interaction objects for mobile resources. Simulation results show that the proposed MARL scheme outperforms common schemes of 3GPP LTE, traditional reinforcement learning schemes, and random selection schemes in improving communication rate and stability.
引用
收藏
页码:3943 / 3954
页数:12
相关论文
共 25 条
[1]  
[Anonymous], 2013, Tech. Rep. 36.819
[2]   Aerial Remote Sensing in Agriculture: A Practical Approach to Area Coverage and Path Planning for Fleets of Mini Aerial Robots [J].
Barrientos, Antonio ;
Colorado, Julian ;
del Cerro, Jaime ;
Martinez, Alexander ;
Rossi, Claudio ;
Sanz, David ;
Valente, Joao .
JOURNAL OF FIELD ROBOTICS, 2011, 28 (05) :667-689
[3]   Infinite-horizon policy-gradient estimation [J].
Baxter, J ;
Bartlett, PL .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2001, 15 :319-350
[4]  
Ding W., 2021, P IEEE 13 INT C WIR, P1
[5]   Towards Multi-agent Reinforcement Learning for Wireless Network Protocol Synthesis [J].
Dutta, Hrishikesh ;
Biswas, Subir .
2021 INTERNATIONAL CONFERENCE ON COMMUNICATION SYSTEMS & NETWORKS (COMSNETS), 2021, :614-622
[6]   Deep Q-Networks for Aerial Data Collection in Multi-UAV-Assisted Wireless Sensor Networks [J].
Emami, Yousef ;
Wei, Bo ;
Li, Kai ;
Ni, Wei ;
Tovar, Eduardo .
IWCMC 2021: 2021 17TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE (IWCMC), 2021, :669-674
[7]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[8]  
Ibrahim A. M., 2022, P IEEE 6 INT S TEL T, P62
[9]  
Iturria-rivera Pedro Enrique, 2022, 2022 IEEE 19th Annual Consumer Communications & Networking Conference (CCNC), P796, DOI 10.1109/CCNC49033.2022.9700667
[10]   Lyapunov Optimization Based Mobile Edge Computing for Internet of Vehicles Systems [J].
Jia, Yi ;
Zhang, Cheng ;
Huang, Yongming ;
Zhang, Wei .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2022, 70 (11) :7418-7433