Energy-Efficient Trajectory Optimization With Wireless Charging in UAV-Assisted MEC Based on Multi-Objective Reinforcement Learning

被引:7
|
作者
Song, Fuhong [1 ]
Deng, Mingsen [1 ]
Xing, Huanlai [2 ]
Liu, Yanping [3 ]
Ye, Fei [4 ]
Xiao, Zhiwen [2 ]
机构
[1] Guizhou Univ Finance & Econ, Sch Informat, Guiyang 550025, Peoples R China
[2] Southwest Jiaotong Univ, Sch Comp & Artificial Intelligence, Chengdu 611756, Peoples R China
[3] Guizhou Univ Finance & Econ, Coll Big Data Stat, Guiyang 550025, Peoples R China
[4] Univ York, Dept Comp Sci, York YO10 5GH, England
基金
中国国家自然科学基金;
关键词
Autonomous aerial vehicles; Task analysis; Energy efficiency; Laser beams; Heuristic algorithms; Reinforcement learning; Inductive charging; Mobile edge computing; multi-objective reinforcement learning; trajectory optimization; unmanned aerial vehicle; wireless charging; ALLOCATION; TASK; CONSUMPTION; ALGORITHM;
D O I
10.1109/TMC.2024.3384405
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper investigates the problem of energy-efficient trajectory optimization with wireless charging (ETWC) in an unmanned aerial vehicle (UAV)-assisted mobile edge computing system. A UAV is dispatched to collect computation tasks from specific ground smart devices (GSDs) within its coverage while transmitting energy to the other GSDs. In addition, a high-altitude platform with a laser beam is deployed in the stratosphere to charge the UAV, so as to maintain its flight mission. The ETWC problem is characterized by multi-objective optimization, aiming to maximize both the energy efficiency of the UAV and the number of tasks collected via optimizing the UAV's flight trajectories. The conflict between the two objectives in the problem makes it quite challenging. Recently, some single-objective reinforcement learning (SORL) algorithms have been introduced to address the aforementioned problem. Nevertheless, these SORLs adopt linear scalarization to define the user utility, thus ignoring the conflict between objectives. Furthermore, in dynamic MEC scenarios, the relative importance assigned to each objective may vary over time, posing significant challenges for conventional SORLs. To solve the challenge, we first build a multi-objective Markov decision process that has a vectorial reward mechanism. There is a corresponding relationship between each component of the reward and one of the two objectives. Then, we propose a new trace-based experience replay scheme to modify sample efficiency and reduce replay buffer bias, resulting in a modified multi-objective reinforcement learning algorithm. The experiment results validate that the proposed algorithm can obtain better adaptability to dynamic preferences and a more favorable balance between objectives compared with several algorithms.
引用
收藏
页码:10867 / 10884
页数:18
相关论文
共 50 条
  • [1] Evolutionary Multi-Objective Reinforcement Learning Based Trajectory Control and Task Offloading in UAV-Assisted Mobile Edge Computing
    Song, Fuhong
    Xing, Huanlai
    Wang, Xinhan
    Luo, Shouxi
    Dai, Penglin
    Xiao, Zhiwen
    Zhao, Bowen
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (12) : 7387 - 7405
  • [2] Dynamic Trajectory Design and Bandwidth Adjustment for Energy-Efficient UAV-Assisted Relaying With Deep Reinforcement Learning in MEC IoT System
    Du, Tianjiao
    Gui, Xiaolin
    Teng, Xiaoyu
    Zhang, Kaiyuan
    Ren, Dewang
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (23): : 37463 - 37479
  • [3] A Spiking Reinforcement Trajectory Planning for UAV-Assisted MEC Systems
    Xia, Zeyang
    Dong, Li
    Jiang, Feibo
    IEEE ACCESS, 2024, 12 : 54435 - 54448
  • [4] Energy-Efficient Trajectory Optimization for UAV-Assisted IoT Networks
    Zhang, Liang
    Celik, Abdulkadir
    Dang, Shuping
    Shihada, Basem
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2022, 21 (12) : 4323 - 4337
  • [5] Learning-Based Resource Management Optimization for UAV-Assisted MEC Against Jamming
    Liu, Shuai
    Yang, Helin
    Xiao, Liang
    Zheng, Mengting
    Lu, Huabing
    Xiong, Zehui
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2024, 72 (08) : 4873 - 4886
  • [6] Reinforcement Learning for Energy-Efficient User Association in UAV-Assisted Cellular Networks
    Kaleem, Zeeshan
    Khalid, Waqas
    Ahmad, Ayaz
    Yu, Heejung
    Almasoud, Abdullah M.
    Yuen, Chau
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2024, 60 (02) : 2474 - 2481
  • [7] Deep Reinforcement Learning-Empowered Trajectory and Resource Allocation Optimization for UAV-Assisted MEC Systems
    Sun, Haowen
    Chen, Ming
    Pan, Yijin
    Cang, Yihan
    Zhao, Jiahui
    Sun, Yuanzhi
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2024, 13 (07) : 1823 - 1827
  • [8] Multi-Objective Deployment Optimization of UAVs for Energy-Efficient Wireless Coverage
    Zhu, Xiumin
    Zhai, Linbo
    Li, Nianxin
    Li, Yumei
    Yang, Feng
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2024, 72 (06) : 3587 - 3601
  • [9] A Reinforcement Learning-Based Stochastic Game for Energy-Efficient UAV Swarm-Assisted MEC With Dynamic Clustering and Scheduling
    Li, Jialiuyuan
    Yi, Changyan
    Chen, Jiayuan
    Shi, You
    Zhang, Tong
    Li, Xiaolong
    Wang, Ran
    Zhu, Kun
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2025, 9 (01): : 255 - 270
  • [10] Distributed Optimization of Multi-Role UAV Functionality Switching and Trajectory for Security Task Offloading in UAV-Assisted MEC
    Zhong, Liang
    Liu, Yuyang
    Deng, Xianjun
    Wu, Celimuge
    Liu, Shenghao
    Yang, Laurence T.
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (12) : 19432 - 19447