MO-AVC: Deep-Reinforcement-Learning-Based Trajectory Control and Task Offloading in Multi-UAV-Enabled MEC Systems

被引:3
|
作者
Gao, Zhen [1 ]
Yang, Lei [1 ]
Dai, Yu [2 ]
机构
[1] Northeastern Univ, Sch Comp Sci & Engn, Shenyang 110000, Peoples R China
[2] Northeastern Univ, Sch Coll Software, Shenyang 110000, Peoples R China
关键词
Task analysis; Optimization; Training; Reinforcement learning; Internet of Things; Autonomous aerial vehicles; Vehicle dynamics; Mobile edge computing (MEC); multiobjective reinforcement learning (MORL); task offloading; trajectory control; unmanned aerial vehicle (UAV); RESOURCE-ALLOCATION; IOT NETWORKS; COMMUNICATION;
D O I
10.1109/JIOT.2023.3329869
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We investigate the joint trajectory control and task offloading (JTCTO) problem in multiunmanned aerial vehicle (UAV)-enabled mobile edge computing (MEC). However, existing JTCTO solutions primarily focus on fixed UAV-enabled MEC scenario variations and necessitate extensive interaction to adapt to new scenarios. Moreover, we consider minimizing task latency and UAV's energy consumption, and maximizing the quantity of tasks collected by the UAV as optimization goals. However, this optimization problem is characterized by multiple conflicting goals that should be adjusted according to their relative significance. In this article, we present a multiobjective actor-variations critic-based JTCTO solution (MO-AVC). First, a group of reinforcement learning strategies is utilized to collect experience on training scenarios, which are employed to learn embeddings of both strategies and scenarios. Further, these two embeddings are used as inputs to train the actor-variations critic (AVC), which explicitly estimates the total return in a space of JTCTO strategies and UAV-enabled MEC scenarios. When adapting to a new scenario, just a few steps of scenario interaction are enough to predict the scenario embedding, thus selecting strategies by maximizing the trained AVC. Second, we propose an actor-conditioned critic framework where the outputs are conditioned on the varying significance of goals, and present a weight dynamic memory-based experience replay to address the intrinsic instability of the dynamic weight context. Finally, simulation results show that MO-AVC can quickly adapt to new scenarios. Moreover, MO-AVC reduces the latency by 7.56%-10.57%, the energy consumption by 11.11%-17.27%, and increases the tasks number by 10.33%-15.54% compared to existing solutions.
引用
收藏
页码:11395 / 11414
页数:20
相关论文
共 50 条
  • [1] Secure Video Offloading in Multi-UAV-Enabled MEC Networks: A Deep Reinforcement Learning Approach
    Zhao, Tantan
    Li, Fan
    He, Lijun
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (02) : 2950 - 2963
  • [2] Computation Offloading and Trajectory Planning of Multi-UAV-Enabled MEC: A Knowledge-Assisted Multiagent Reinforcement Learning Approach
    Li, Xulong
    Qin, Yunhui
    Huo, Jiahao
    Wei, Huangfu
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (05) : 7077 - 7088
  • [3] Deep Reinforcement Learning Based 3D-Trajectory Design and Task Offloading in UAV-Enabled MEC System
    Liu, Chuanjie
    Zhong, Yalin
    Wu, Ruolin
    Ren, Siyu
    Du, Shuang
    Guo, Bing
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2025, 74 (02) : 3185 - 3195
  • [4] Joint computation offloading and deployment optimization in multi-UAV-enabled MEC systems
    Chen, Zheyi
    Zheng, Hongqiang
    Zhang, Jianshan
    Zheng, Xianghan
    Rong, Chunming
    PEER-TO-PEER NETWORKING AND APPLICATIONS, 2022, 15 (01) : 194 - 205
  • [5] Joint computation offloading and deployment optimization in multi-UAV-enabled MEC systems
    Zheyi Chen
    Hongqiang Zheng
    Jianshan Zhang
    Xianghan Zheng
    Chunming Rong
    Peer-to-Peer Networking and Applications, 2022, 15 : 194 - 205
  • [6] Deep reinforcement learning based trajectory design and resource allocation for task-aware multi-UAV enabled MEC networks
    Li, Zewu
    Xu, Chen
    Zhang, Zhanpeng
    Wu, Runze
    COMPUTER COMMUNICATIONS, 2024, 213 : 88 - 98
  • [7] DEEP REINFORCEMENT LEARNING FOR COMPUTATION OFFLOADING AND RESOURCE ALLOCATION IN BLOCKCHAIN-BASED MULTI-UAV-ENABLED MOBILE EDGE COMPUTING
    Mohammed, Abegaz
    Nahom, Hayla
    Tewodros, Ayall
    Habtamu, Yasin
    Hayelow, Gebrye
    2020 17TH INTERNATIONAL COMPUTER CONFERENCE ON WAVELET ACTIVE MEDIA TECHNOLOGY AND INFORMATION PROCESSING (ICCWAMTIP), 2020, : 295 - 299
  • [8] A Novel Deep Reinforcement Learning Approach for Task Offloading in MEC Systems
    Liu, Xiaowei
    Jiang, Shuwen
    Wu, Yi
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [9] Multiobjective Deep Reinforcement Learning for Computation Offloading and Trajectory Control in UAV-Base-Station-Assisted MEC
    Huang, Hao
    Chai, Zheng-Yi
    Sun, Bao-Shan
    Kang, Hong-Shen
    Zhao, Ying-Jie
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 31805 - 31821
  • [10] Asynchronous Federated Deep-Reinforcement-Learning-Based Dependency Task Offloading for UAV-Assisted Vehicular Networks
    Shen, Si
    Shen, Guojiang
    Dai, Zhehao
    Zhang, Kaiyu
    Kong, Xiangjie
    Li, Jianxin
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (19): : 31561 - 31574