Meta-Hierarchical Reinforcement Learning (MHRL)-Based Dynamic Resource Allocation for Dynamic Vehicular Networks

被引:49
作者
He, Ying [1 ]
Wang, Yuhang [1 ]
Lin, Qiuzhen [1 ]
Li, Jianqiang [1 ]
机构
[1] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen 518060, Peoples R China
关键词
Reinforcement learning; Task analysis; Vehicle dynamics; Resource management; Dynamic scheduling; Training; Adaptation models; Dynamic vehicular networks; hierarchical reinforcement learning; meta-learning; resource allocation; INFORMATION-CENTRIC NETWORKING; CONNECTED VEHICLES; FRAMEWORK;
D O I
10.1109/TVT.2022.3146439
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the rapid development of vehicular networks, there is an increasing demand for extensive networking, computting, and caching resources. How to allocate multiple resources effectively and efficiently for dynamic vehicular networks is extremely important. Most existing works on resource management in vehicular networks assume static network conditions. In this paper, we propose a general framework that can enable fast-adaptive resource allocation for dynamic vehicular environments. Specifically, we model the dynamics of the vehicular environment as a series of related Markov Decision Processes (MDPs), and we combine hierarchical reinforcement learning with meta-learning, which makes our proposed framework quickly adapt to a new environment by only fine-tuning the top-level master network, and meanwhile the low-level sub-networks can make the right resource allocation policy. Extensive simulation results show the effectiveness of our proposed framework, which can quickly adapt to different scenarios, and significantly improve the performance of resource management in dynamic vehicular networks.
引用
收藏
页码:3495 / 3506
页数:12
相关论文
共 43 条
[1]   Information-Centric Networking for Connected Vehicles: A Survey and Future Perspectives [J].
Amadeo, Marica ;
Campolo, Claudia ;
Molinaro, Antonella .
IEEE COMMUNICATIONS MAGAZINE, 2016, 54 (02) :98-104
[2]  
[Anonymous], 2010, Synthesis Lectures on Artificial Intelligence and Machine Learning
[3]   A View of Cloud Computing [J].
Armbrust, Michael ;
Fox, Armando ;
Griffith, Rean ;
Joseph, Anthony D. ;
Katz, Randy ;
Konwinski, Andy ;
Lee, Gunho ;
Patterson, David ;
Rabkin, Ariel ;
Stoica, Ion ;
Zaharia, Matei .
COMMUNICATIONS OF THE ACM, 2010, 53 (04) :50-58
[4]  
Barto AG, 2003, DISCRETE EVENT DYN S, V13, P41, DOI [10.1023/A:1022140919877, 10.1023/A:1025696116075]
[5]   Edge Computing Resources Reservation in Vehicular Networks: A Meta-Learning Approach [J].
Chen, Dawei ;
Liu, Yin-Chen ;
Kim, BaekGyu ;
Xie, Jiang ;
Hong, Choong Seon ;
Han, Zhu .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (05) :5634-5646
[6]   Dynamical Control Domain Division for Software-Defined Satellite-Ground Integrated Vehicular Networks [J].
Chen, Long ;
Tang, Feilong ;
Li, Xu ;
Yang, Laurence T. ;
Cao, Lijun ;
Yu, Jiadi ;
Fu, Luoyi ;
Li, Zhetao ;
Kong, Linghe .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2021, 8 (04) :2732-2741
[7]   Hierarchical reinforcement learning with the MAXQ value function decomposition [J].
Dietterich, TG .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2000, 13 :227-303
[8]   Computation Offloading and Resource Allocation in Vehicular Networks Based on Dual-Side Cost Minimization [J].
Du, Jianbo ;
Yu, F. Richard ;
Chu, Xiaoli ;
Feng, Jie ;
Lu, Guangyue .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (02) :1079-1092
[9]  
Finn C, 2017, PR MACH LEARN RES, V70
[10]  
Frans K., 2018, INT C LEARNING REPRE