MSRA: Mode Selection and Resource Allocation for Cooperative Vehicle-Infrastructure System

被引:0
作者
Tian, Jin [1 ]
Shi, Yan [1 ]
Xu, Yaqi [1 ]
Chen, Shanzhi [2 ,3 ]
Ge, Yuming [4 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[2] China Acad Telecommun Technol, Natl Engn Res Ctr Mobile Commun & Vehicular Networ, Beijing 100191, Peoples R China
[3] China Acad Telecommun Technol, State Key Lab Wireless Mobile Commun, Beijing 100191, Peoples R China
[4] Minist Ind & Informat Technol, Key Lab Internet Vehicle Tech Innovat & Testing CA, Beijing 100191, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Resource management; Quality of service; Reliability; Vehicle dynamics; Reliability engineering; Reinforcement learning; Optimization; Low latency communication; Buildings; Vehicular ad hoc networks; Cellular vehicle to everything (C-V2X); cooperative vehicle-infrastructure systems (CVISs); mode selection; multiagent reinforcement learning (MARL); resource allocation; CHALLENGES; TECHNOLOGIES; INTERNET;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Cellular vehicle-to-everything (C-V2X) communication is essential for supporting diverse vehicle applications in cooperative vehicle-infrastructure systems (CVISs). However, various factors, such as building obstructions, signal blockages, and network conditions, severely affect the latency and reliability of V2V communication. To address this challenge, we propose a novel dynamic method for communication mode selection and resource allocation in C-V2X, termed MSRA, which enables vehicles to dynamically select and utilize different communication modes, such as V2V, Vehicle-to-Infrastructure (V2I), and Vehicle-to-Network (V2N) for direct, relay, or forwarding communication, based on real-time conditions. To this end, we formulate the joint optimization problem of mode selection and resource allocation as a Markov decision process (MDP) and propose a solution based on multiagent reinforcement learning (MARL) with centralized training and decentralized execution. Specifically, each vehicle acts as an agent, independently selecting communication modes and resources based on real-time network status and communication link quality, aiming to satisfy the latency and reliability requirements of V2V communication while maximizing the capacity of V2N communication based on heterogeneous Quality-of-Service (QoS) requirements. Simulation results indicate that the proposed algorithm significantly outperforms other decentralized baselines and demonstrates superior performance under various conditions.
引用
收藏
页码:9927 / 9939
页数:13
相关论文
共 43 条
[21]   Multi-Agent Reinforcement Learning Resources Allocation Method Using Dueling Double Deep Q-Network in Vehicular Networks [J].
Ji, Yuxin ;
Wang, Yu ;
Zhao, Haitao ;
Gui, Guan ;
Gacanin, Haris ;
Sari, Hikmet ;
Adachi, Fumiyuki .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (10) :13447-13460
[22]   Resource Allocation for D2D-Based V2X Communication With Imperfect CSI [J].
Li, Xiaoshuai ;
Ma, Lin ;
Xu, Yubin ;
Shankaran, Rajan .
IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (04) :3545-3558
[23]   Graph-Based Resource Sharing in Vehicular Communication [J].
Liang, Le ;
Xie, Shijie ;
Li, Geoffrey Ye ;
Ding, Zhi ;
Yu, Xingxing .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (07) :4579-4592
[24]   Spectrum and Power Allocation for Vehicular Communications With Delayed CSI Feedback [J].
Liang, Le ;
Kim, Joonbeom ;
Jha, Satish C. ;
Sivanesan, Kathiravetpillai ;
Li, Geoffrey Ye .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2017, 6 (04) :458-461
[25]   Ultra-Reliable and Low-Latency Vehicular Transmission: An Extreme Value Theory Approach [J].
Liu, Chen-Feng ;
Bennis, Mehdi .
IEEE COMMUNICATIONS LETTERS, 2018, 22 (06) :1292-1295
[26]   Multitimescale Control and Communications With Deep Reinforcement Learning-Part I: Communication-Aware Vehicle Control [J].
Liu, Tong ;
Lei, Lei ;
Zheng, Kan ;
Shen, Xuemin .
IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (09) :15386-15401
[27]   Conformal Metasurfaces: A Novel Solution for Vehicular Communications [J].
Mizmizi, Marouan ;
Ayoubi, Reza Aghazadeh ;
Tagliaferri, Dario ;
Dong, Kai ;
Gentili, Gian Guido ;
Spagnolini, Umberto .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (04) :2804-2817
[28]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533
[29]   Edge-Enabled V2X Service Placement for Intelligent Transportation Systems [J].
Moubayed, Abdallah ;
Shami, Abdallah ;
Heidari, Parisa ;
Larabi, Adel ;
Brunner, Richard .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2021, 20 (04) :1380-1392
[30]   A Survey on Resource Allocation in Vehicular Networks [J].
Noor-A-Rahim, Md ;
Liu, Zilong ;
Lee, Haeyoung ;
Ali, G. G. Md Nawaz ;
Pesch, Dirk ;
Xiao, Pei .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (02) :701-721