Joint Service Migration and Resource Allocation in Edge IoT System Based on Deep Reinforcement Learning

被引:15
作者
Liu, Fangzheng [1 ]
Yu, Hao [2 ]
Huang, Jiwei [1 ]
Taleb, Tarik [2 ]
机构
[1] China Univ Petr, Beijing Key Lab Petr Data Min, Beijing 102249, Peoples R China
[2] Univ Oulu, Ctr Wireless Commun, Oulu, Finland
基金
中国国家自然科学基金;
关键词
Resource management; Servers; Internet of Things; Optimization; Delays; Heuristic algorithms; Vehicle dynamics; Cloud; deep reinforcement learning (DRL); edge computing; Internet of Things (IoT); long short term memory (LSTM); multiaccess edge computing (MEC); parameterized deep Q-networks (PDQNs); resource allocation; service migration; NETWORKS; CLOUD;
D O I
10.1109/JIOT.2023.3332421
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multiaccess edge computing (MEC) provides services for resource-sensitive and delay-sensitive Internet of Things (IoT) applications by extending the capabilities of cloud computing to the edge of the networks. However, the high mobility of IoT devices (e.g., vehicles) and the limited resources of edge servers (ESs) affect the service continuity and access latency. Service migration and reasonable resource (re-)allocation consequently become needed to ensure Quality of Service (QoS). However, service migration results in additional latency. In addition, different mobile IoT users have different resource requirements and different resource allocation policies of target ESs also determine whether service migration is necessary. Subsequently, how to jointly optimize service migration and resource allocation (SMRA) is a challenge that needs to be carefully addressed. To this end, this article investigates the joint optimization problem of SMRA in MEC environments to minimize the access delay of IoT users. It proposes a joint SMRA algorithm based on deep reinforcement learning (DRL), which takes into account the mobility of IoT users and decides whether to migrate services, where to migrate, and how to allocate resources through the long short time memory (LSTM) algorithm and the parameterized deep $Q$ -network (PDQN) algorithm. Moreover, the PDQN algorithm effectively solves the discrete-continuous hybrid action space challenge in the SMRA problem. Finally, we conduct evaluation using a real-world data set of Beijing cab trajectories to verify the effectiveness and superiority of our proposed SMRA solution.
引用
收藏
页码:11341 / 11352
页数:12
相关论文
共 28 条
[1]   AI-Based Network-Aware Service Function Chain Migration in 5G and Beyond Networks [J].
Addad, Rami Akrem ;
Dutra, Diego Leonel Cadette ;
Taleb, Tarik ;
Flinck, Hannu .
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2022, 19 (01) :472-484
[2]   Toward Using Reinforcement Learning for Trigger Selection in Network Slice Mobility [J].
Addad, Rami Akrem ;
Cadette Dutra, Diego Leonel ;
Taleb, Tarik ;
Flinck, Hannu .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (07) :2241-2253
[3]  
[Anonymous], 2013, Rep. TS 36.331 V8.20.0
[4]   Efficient Multi-User Computation Offloading for Mobile-Edge Cloud Computing [J].
Chen, Xu ;
Jiao, Lei ;
Li, Wenzhong ;
Fu, Xiaoming .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2016, 24 (05) :2827-2840
[5]   Dynamic Task Allocation and Service Migration in Edge-Cloud IoT System Based on Deep Reinforcement Learning [J].
Chen, Yan ;
Sun, Yanjing ;
Wang, Chenyang ;
Taleb, Tarik .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (18) :16742-16757
[6]   Edge Intelligence for Energy-Efficient Computation Offloading and Resource Allocation in 5G Beyond [J].
Dai, Yueyue ;
Zhang, Ke ;
Maharjan, Sabita ;
Zhang, Yan .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (10) :12175-12186
[7]   Deep reinforcement learning control of electric vehicle charging in the of [J].
Dorokhova, Marina ;
Martinson, Yann ;
Ballif, Christophe ;
Wyrsch, Nicolas .
APPLIED ENERGY, 2021, 301
[8]  
Hochreiter S, 1997, NEURAL COMPUT, V9, P1735, DOI [10.1162/neco.1997.9.8.1735, 10.1162/neco.1997.9.1.1, 10.1007/978-3-642-24797-2]
[9]  
Horii M, 2018, IEEE WCNC
[10]  
Jiajia Liu, 2015, 2015 IEEE Conference on Computer Communications (INFOCOM). Proceedings, P46, DOI 10.1109/INFOCOM.2015.7218366