MARS: A DRL-Based Multi-Task Resource Scheduling Framework for UAV With IRS-Assisted Mobile Edge Computing System

被引:11
作者
Jiang, Feibo [1 ]
Peng, Yubo [2 ]
Wang, Kezhi [3 ]
Dong, Li [4 ]
Yang, Kun [5 ,6 ]
机构
[1] Hunan Normal Univ, Hunan Prov Key Lab Intelligent Comp & Language In, Changsha 410081, Peoples R China
[2] Hunan Normal Univ, Sch Informat Sci & Engn, Changsha 410081, Peoples R China
[3] Brunel Univ London, Dept Comp Sci, Uxbridge UB8 3PH, England
[4] Hunan Univ Technol & Business, Xiangjiang Lab, Changsha 410205, Peoples R China
[5] Univ Essex, Sch Comp Sci & Elect Engn, Colchester CO4 3SQ, England
[6] Changchun Inst Technol, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Mobile edge computing (MEC); intelligent reflecting surface (IRS); unmanned aerial vehicle (UAV); deep reinforcement learning (DRL); resource scheduling; ALLOCATION; OPTIMIZATION; DESIGN;
D O I
10.1109/TCC.2023.3307582
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies a dynamic Mobile Edge Computing (MEC) system assisted by Unmanned Aerial Vehicles (UAVs) and Intelligent Reflective Surfaces (IRSs). We propose a scaleable resource scheduling algorithm to minimize the energy consumption of all UEs and UAVs in the MEC system with a variable number of UAVs. We propose a Multi-tAsk Resource Scheduling (MARS) framework based on Deep Reinforcement Learning (DRL) to solve the problem. First, we present a novel Advantage Actor-Critic (A2C) structure with the state-value critic and entropy-enhanced actor to reduce variance and enhance the policy search of DRL. Then, we present a multi-head agent with three different heads in which a classification head is applied to make offloading decisions and a regression head is presented to allocate computational resources, and a critic head is introduced to estimate the state value of the selected action. Next, we introduce a multi-task controller to adjust the agent to adapt to the varying number of UAVs by loading or unloading a part of weights in the agent. Finally, a Light Wolf Search (LWS) is introduced as the action refinement to enhance the exploration in the dynamic action space. The numerical results demonstrate the feasibility and efficiency of the MARS framework.
引用
收藏
页码:3700 / 3712
页数:13
相关论文
共 41 条
[1]  
Bhandari J., 2018, C LEARN THEOR, V75, P1691
[2]   Resource Allocation for URLLC-Oriented Two-Way UAV Relaying [J].
Cai, Yeming ;
Jiang, Xu ;
Liu, Mingqian ;
Zhao, Nan ;
Chen, Yunfei ;
Wang, Xianbin .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (03) :3344-3349
[3]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[4]   Energy Efficient Resource Allocation for IRS Assisted CoMP Systems [J].
Chen, Jian ;
Xie, Yunhe ;
Mu, Xidong ;
Jia, Jie ;
Liu, Yuanwei ;
Wang, Xingwei .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (07) :5688-5702
[5]   Resource Allocation for IRS Assisted SGF NOMA Transmission: A MADRL Approach [J].
Chen, Jian ;
Guo, Liang ;
Jia, Jie ;
Shang, Jianhui ;
Wang, Xingwei .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2022, 40 (04) :1302-1316
[6]   TABOO SEARCH - AN APPROACH TO THE MULTIPLE MINIMA PROBLEM [J].
CVIJOVIC, D ;
KLINOWSKI, J .
SCIENCE, 1995, 267 (5198) :664-666
[7]   A Multi-Agent Collaborative Environment Learning Method for UAV Deployment and Resource Allocation [J].
Dai, Zhaojun ;
Zhang, Yan ;
Zhang, Wancheng ;
Luo, Xinran ;
He, Zunwen .
IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2022, 8 (120-130) :120-130
[8]  
Du Y, 2018, IEEE GLOB COMM CONF
[9]  
Dulac-Arnold G., 2015, ARXIV
[10]  
Fujimoto S, 2018, PR MACH LEARN RES, V80