Computation Migration and Resource Allocation in Heterogeneous Vehicular Networks: A Deep Reinforcement Learning Approach

被引:18
作者
Wang, Hui [1 ]
Ke, Hongchang [2 ,3 ,4 ]
Liu, Gang [1 ]
Sun, Weijia [1 ]
机构
[1] Changchun Univ Technol, Coll Comp Sci & Engn, Changchun 130012, Peoples R China
[2] Jilin Univ, Coll Comp Sci & Technol, Changchun 130012, Peoples R China
[3] Changchun Inst Technol, Sch Comp Technol & Engn, Changchun 130012, Peoples R China
[4] Jilin Univ, Key Lab Symbol Computat & Knowledge Engn, Minist Educ, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Servers; Resource management; Task analysis; Delays; Computational modeling; Base stations; Edge computing; Vehicular networks; mobile edge computing; reinforcement learning; computation migration; MOBILE; AWARE; MEC;
D O I
10.1109/ACCESS.2020.3024683
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the development of 5G technology, the requirements for data communication and computation in emerging 5G-enabled vehicular networks are becoming increasingly stringent. Computation-intensive or delay-sensitive tasks generated by vehicles need to be processed in real time. Mobile edge computing (MEC) is an appropriate solution. Wireless users or vehicles can offload computation tasks to the MEC server due to it has strong computation ability and is closer to the wireless users or vehicles. However, the communication and computation resources of the single MEC are not sufficient for executing the continuously generated computation-intensive or delay-sensitive tasks. We consider migrating computation tasks to other MEC servers to reduce the computation and communication pressure on current MEC server. In this article, we construct an MEC-based computation offloading framework for vehicular networks, which considers time-varying channel states and stochastically arriving computation tasks. To minimize the total cost of the proposed MEC framework, which consists of the delay cost, energy computation cost, and bandwidth cost, we propose a deep reinforcement learning-based computation migration and resource allocation (RLCMRA) scheme that requires no prior knowledge. The RLCMRA algorithm can obtain the optimal offloading and migration policy by adaptive learning to maximize the average cumulative reward (minimize the total cost). Extensive numerical results show that the proposed RLCMRA algorithm can adaptively learn the optimal policy and outperform four other baseline algorithms.
引用
收藏
页码:171140 / 171153
页数:14
相关论文
共 44 条
[31]  
Singhal C, 2017, ADV WIREL TECHNOL TE, P1, DOI 10.4018/978-1-5225-2023-8
[32]   Effect of Feedback Delay on Amplify-and-Forward Relay Networks With Beamforming [J].
Suraweera, Himal A. ;
Tsiftsis, Theodoros A. ;
Karagiannidis, George K. ;
Nallanathan, Arumugam .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2011, 60 (03) :1265-1271
[33]  
Sutton R. S., 2020, P ADV NEUR INF PROC, V12, P1057
[34]  
Sutton R. S., 2020, Reinforcement Learning, An Introduction, V2nd
[35]  
Tsiropoulou EE, 2014, LECT NOTES COMPUT SC, V8487, P114
[36]  
van Hasselt H, 2016, AAAI CONF ARTIF INTE, P2094
[37]   Joint Computation Offloading and Interference Management in Wireless Cellular Networks with Mobile Edge Computing [J].
Wang, Chenmeng ;
Yu, F. Richard ;
Liang, Chengchao ;
Chen, Qianbin ;
Tang, Lun .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (08) :7432-7445
[38]   Computation Offloading in Multi-Access Edge Computing Using a Deep Sequential Model Based on Reinforcement Learning [J].
Wang, Jin ;
Hu, Jia ;
Min, Geyong ;
Zhan, Wenhan ;
Ni, Qiang ;
Georgalas, Nektarios .
IEEE COMMUNICATIONS MAGAZINE, 2019, 57 (05) :64-69
[39]   Delay-Aware Microservice Coordination in Mobile Edge Computing: A Reinforcement Learning Approach [J].
Wang, Shangguang ;
Guo, Yan ;
Zhang, Ning ;
Yang, Peng ;
Zhou, Ao ;
Shen, Xuemin .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2021, 20 (03) :939-951
[40]   Dynamic Service Placement for Mobile Micro-Clouds with Predicted Future Costs [J].
Wang, Shiqiang ;
Urgaonkar, Rahul ;
He, Ting ;
Chan, Kevin ;
Zafer, Murtaza ;
Leung, Kin K. .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2017, 28 (04) :1002-1016