Multi-objective deep reinforcement learning for computation offloading in UAV-assisted multi-access edge computing ✩

被引:29
作者
Liu, Xu [1 ]
Chai, Zheng-Yi [2 ]
Li, Ya-Lun [3 ]
Cheng, Yan-Yang [2 ]
Zeng, Yue [4 ]
机构
[1] Tiangong Univ, Sch Software, Tianjin 300387, Peoples R China
[2] Tiangong Univ, Sch Comp Sci & Technol, Tianjin 300387, Peoples R China
[3] Tiangong Univ, Sch Elect & Informat Engn, Tianjin 300387, Peoples R China
[4] Jinling Inst Technol, Sch Software Engn, Nanjing 211199, Peoples R China
基金
中国国家自然科学基金;
关键词
Unmanned aerial vehicle; Multi-access edge computing; Computation offloading; Multi-objective; Reinforcement learning; ALGORITHM;
D O I
10.1016/j.ins.2023.119154
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Unmanned aerial vehicle-assisted multi-access edge computing (UAV-MEC) plays an important role in some complex environments such as mountainous and disaster areas. Computation offloading problem (COP) is one of the key issues of UAV-MEC, which mainly aims to minimize the conflict goals between energy consumption and delay. Due to the time-varying and uncertain nature of the UAV-MEC system, deep reinforcement learning is an effective method for solving the COP. Different from the existing works, in this paper, the COP in UAV-MEC system is modeled as a multi-objective Markov decision process, and a multi-objective deep reinforcement learning method is proposed to solve it. In the proposed algorithm, the scalar reward of reinforcement learning is expanded into a vector reward, and the weights are dynamically adjusted to meet different user preferences. The most important preferences are selected by non-dominated sorting, which can better maintain the previously learned strategy. In addition, the Q network structure combines Double Deep Q Network (Double DQN) with Dueling Deep Q Network (Dueling DQN) to improve the optimization efficiency. Simulation results show that the algorithm achieves a good balance between energy consumption and delay, and can obtain a better computation offloading scheme.
引用
收藏
页数:17
相关论文
共 48 条
[1]  
Abels A, 2019, PR MACH LEARN RES, V97
[2]   Energy Efficient UAV-Enabled Mobile Edge Computing for IoT Devices: A Review [J].
Abrar, Muhammad ;
Ajmal, Ushna ;
Almohaimeed, Ziyad M. ;
Gui, Xiang ;
Akram, Rizwan ;
Masroor, Roha .
IEEE ACCESS, 2021, 9 :127779-127798
[3]   Delay-Optimal Task Offloading for UAV-Enabled Edge-Cloud Computing Systems [J].
Almutairi, Jaber ;
Aldossary, Mohammad ;
Alharbi, Hatem A. ;
Yosuf, Barzan A. ;
Elmirghani, Jaafar M. H. .
IEEE ACCESS, 2022, 10 :51575-51586
[4]   Intelligent ubiquitous computing for future UAV-enabled MEC network systems [J].
Chen, Lunyuan ;
Zhao, Rui ;
He, Ke ;
Zhao, Zichao ;
Fan, Liseng .
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2022, 25 (04) :2417-2427
[5]   A fast and elitist multiobjective genetic algorithm: NSGA-II [J].
Deb, K ;
Pratap, A ;
Agarwal, S ;
Meyarivan, T .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2002, 6 (02) :182-197
[6]   An Application Placement Technique for Concurrent IoT Applications in Edge and Fog Computing Environments [J].
Goudarzi, Mohammad ;
Wu, Huaming ;
Palaniswami, Marimuthu ;
Buyya, Rajkumar .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2021, 20 (04) :1298-1311
[7]   UAV-Enhanced Intelligent Offloading for Internet of Things at the Edge [J].
Guo, Hongzhi ;
Liu, Jiajia .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2020, 16 (04) :2737-2746
[8]   Multi-objective deep reinforcement learning for emergency scheduling in a water distribution network [J].
Hu, Chengyu ;
Wang, Qiuming ;
Gong, Wenyin ;
Yan, Xuesong .
MEMETIC COMPUTING, 2022, 14 (02) :211-223
[9]   Security modeling and efficient computation offloading for service workflow in mobile edge computing [J].
Huang, Binbin ;
Li, Zhongjin ;
Tang, Peng ;
Wang, Shangguang ;
Zhao, Jun ;
Hu, Haiyang ;
Li, Wanqing ;
Chang, Victor .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2019, 97 :755-774
[10]  
Huda S.A., 2022, Journal of Network and Computer Applications