Multi-Agent Deep Reinforcement Learning for Task Offloading in UAV-Assisted Mobile Edge Computing

被引:230
作者
Zhao, Nan [1 ]
Ye, Zhiyang [1 ]
Pei, Yiyang [2 ]
Liang, Ying-Chang [3 ,4 ]
Niyato, Dusit [5 ]
机构
[1] Hubei Univ Technol, Hubei Collaborat Innovat Ctr High Efficiency Util, Wuhan 430068, Peoples R China
[2] Singapore Inst Technol, Singapore 138683, Singapore
[3] Univ Elect Sci & Technol China UESTC, Ctr Intelligent Networking & Commun CINC, Chengdu 610056, Peoples R China
[4] Peng Cheng Lab, Shenzhen 518066, Guangdong, Peoples R China
[5] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
基金
新加坡国家研究基金会; 中国国家自然科学基金;
关键词
Task analysis; Resource management; Trajectory; Wireless communication; Servers; Optimization; Manganese; Mobile edge computing; UAV networks; task offloading; cooperative offloading; deep reinforcement learning; RESOURCE-ALLOCATION; TRAJECTORY DESIGN; MEC;
D O I
10.1109/TWC.2022.3153316
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Mobile edge computing can effectively reduce service latency and improve service quality by offloading computation-intensive tasks to the edges of wireless networks. Due to the characteristic of flexible deployment, wide coverage and reliable wireless communication, unmanned aerial vehicles (UAVs) have been employed as assisted edge clouds (ECs) for large-scale sparely-distributed user equipment. Considering the limited computation and energy capacities of UAVs, a collaborative mobile edge computing system with multiple UAVs and multiple ECs is investigated in this paper. The task offloading issue is addressed to minimize the sum of execution delays and energy consumptions by jointly designing the trajectories, computation task allocation, and communication resource management of UAVs. Moreover, to solve the above non-convex optimization problem, a Markov decision process is formulated for the multi-UAV assisted mobile edge computing system. To obtain the joint strategy of trajectory design, task allocation, and power management, a cooperative multi-agent deep reinforcement learning framework is investigated. Considering the high-dimensional continuous action space, the twin delayed deep deterministic policy gradient algorithm is exploited. The evaluation results demonstrate that our multi-UAV multi-EC task offloading method can achieve better performance compared with the other optimization approaches.
引用
收藏
页码:6949 / 6960
页数:12
相关论文
共 36 条
[31]   Dynamic Controller Assignment in Software Defined Internet of Vehicles Through Multi-Agent Deep Reinforcement Learning [J].
Yuan, Tingting ;
Neto, Wilson da Rocha ;
Rothenberg, Christian Esteve ;
Obraczka, Katia ;
Barakat, Chadi ;
Turletti, Thierry .
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (01) :585-596
[32]   Multi-UAV-Enabled Mobile-Edge Computing for Time-Constrained IoT Applications [J].
Zhan, Cheng ;
Hu, Han ;
Liu, Zhi ;
Wang, Zhi ;
Mao, Shiwen .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (20) :15553-15567
[33]   Stochastic Computation Offloading and Trajectory Scheduling for UAV-Assisted Mobile Edge Computing [J].
Zhang, Jiao ;
Zhou, Li ;
Tang, Qi ;
Ngai, Edith C. -H. ;
Hu, Xiping ;
Zhao, Haitao ;
Wei, Jibo .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (02) :3688-3699
[34]   Multi-Agent Deep Reinforcement Learning for Trajectory Design and Power Allocation in Multi-UAV Networks [J].
Zhao, Nan ;
Liu, Zehua ;
Cheng, Yiqiang .
IEEE ACCESS, 2020, 8 :139670-139679
[35]   Deep Reinforcement Learning for User Association and Resource Allocation in Heterogeneous Cellular Networks [J].
Zhao, Nan ;
Liang, Ying-Chang ;
Niyato, Dusit ;
Pei, Yiyang ;
Wu, Minghu ;
Jiang, Yunhao .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2019, 18 (11) :5141-5152
[36]   Learning-Based Computation Offloading Approaches in UAVs-Assisted Edge Computing [J].
Zhu, Shichao ;
Gui, Lin ;
Zhao, Dongmei ;
Cheng, Nan ;
Zhang, Qi ;
Lang, Xiupu .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2021, 70 (01) :928-944