MO-AVC: Deep-Reinforcement-Learning-Based Trajectory Control and Task Offloading in Multi-UAV-Enabled MEC Systems

被引:7
作者
Gao, Zhen [1 ]
Yang, Lei [1 ]
Dai, Yu [2 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110000, Peoples R China
[2] Northeastern Univ, Sch Coll Software, Shenyang 110000, Peoples R China
关键词
Task analysis; Optimization; Training; Reinforcement learning; Internet of Things; Autonomous aerial vehicles; Vehicle dynamics; Mobile edge computing (MEC); multiobjective reinforcement learning (MORL); task offloading; trajectory control; unmanned aerial vehicle (UAV); RESOURCE-ALLOCATION; IOT NETWORKS; COMMUNICATION;
D O I
10.1109/JIOT.2023.3329869
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We investigate the joint trajectory control and task offloading (JTCTO) problem in multiunmanned aerial vehicle (UAV)-enabled mobile edge computing (MEC). However, existing JTCTO solutions primarily focus on fixed UAV-enabled MEC scenario variations and necessitate extensive interaction to adapt to new scenarios. Moreover, we consider minimizing task latency and UAV's energy consumption, and maximizing the quantity of tasks collected by the UAV as optimization goals. However, this optimization problem is characterized by multiple conflicting goals that should be adjusted according to their relative significance. In this article, we present a multiobjective actor-variations critic-based JTCTO solution (MO-AVC). First, a group of reinforcement learning strategies is utilized to collect experience on training scenarios, which are employed to learn embeddings of both strategies and scenarios. Further, these two embeddings are used as inputs to train the actor-variations critic (AVC), which explicitly estimates the total return in a space of JTCTO strategies and UAV-enabled MEC scenarios. When adapting to a new scenario, just a few steps of scenario interaction are enough to predict the scenario embedding, thus selecting strategies by maximizing the trained AVC. Second, we propose an actor-conditioned critic framework where the outputs are conditioned on the varying significance of goals, and present a weight dynamic memory-based experience replay to address the intrinsic instability of the dynamic weight context. Finally, simulation results show that MO-AVC can quickly adapt to new scenarios. Moreover, MO-AVC reduces the latency by 7.56%-10.57%, the energy consumption by 11.11%-17.27%, and increases the tasks number by 10.33%-15.54% compared to existing solutions.
引用
收藏
页码:11395 / 11414
页数:20
相关论文
共 65 条
[1]   Optimal LAP Altitude for Maximum Coverage [J].
Al-Hourani, Akram ;
Kandeepan, Sithamparanathan ;
Lardner, Simon .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2014, 3 (06) :569-572
[2]  
[Anonymous], 2017, 3GPP Rep. 36.913
[3]   Data Offloading in UAV-Assisted Multi-Access Edge Computing Systems Under Resource Uncertainty [J].
Apostolopoulos, Pavlos Athanasios ;
Fragkos, Georgios ;
Tsiropoulou, Eirini Eleni ;
Papavassiliou, Symeon .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (01) :175-190
[4]   Multitask Multiobjective Deep Reinforcement Learning-Based Computation Offloading Method for Industrial Internet of Things [J].
Cai, Jun ;
Fu, Hongtian ;
Liu, Yan .
IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (02) :1848-1859
[5]   Joint Multi-Task Offloading and Resource Allocation for Mobile Edge Computing Systems in Satellite IoT [J].
Chai, Furong ;
Zhang, Qi ;
Yao, Haipeng ;
Xin, Xiangjun ;
Gao, Ran ;
Guizani, Mohsen .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (06) :7783-7795
[6]  
Chen X., 2021, PROC IEEE GLOB COMMU, P1
[7]   Distributed Computation Offloading and Trajectory Optimization in Multi-UAV-Enabled Edge Computing [J].
Chen, Xiangyi ;
Bi, Yuanguo ;
Han, Guangjie ;
Zhang, Dongyu ;
Liu, Minghan ;
Shi, Han ;
Zhao, Hai ;
Li, Fengyun .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (20) :20096-20110
[8]   Towards Energy-Efficient Scheduling of UAV and Base Station Hybrid Enabled Mobile Edge Computing [J].
Dai, Bin ;
Niu, Jianwei ;
Ren, Tao ;
Hu, Zheyuan ;
Atiquzzaman, Mohammed .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (01) :915-930
[9]   Multi-Agent Deep Reinforcement Learning for Joint Decoupled User Association and Trajectory Design in Full-Duplex Multi-UAV Networks [J].
Dai, Chen ;
Zhu, Kun ;
Hossain, Ekram .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (10) :6056-6070
[10]   Meta Reinforcement Learning for Multi-Task Offloading in Vehicular Edge Computing [J].
Dai, Penglin ;
Huang, Yaorong ;
Hu, Kaiwen ;
Wu, Xiao ;
Xing, Huanlai ;
Yu, Zhaofei .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (03) :2123-2138