Multi-agent few-shot meta reinforcement learning for trajectory design and channel selection in UAV-assisted networks

被引:9
|
作者
Zhou, Shiyang [1 ]
Cheng, Yufan [1 ]
Lei, Xia [1 ]
Duan, Huanhuan [1 ]
机构
[1] Univ Elect Sci & Technol China, Natl Key Lab Sci & Technol Commun, Chengdu 611731, Peoples R China
关键词
Jamming; Trajectory; Task analysis; Automobiles; Wireless networks; Heuristic algorithms; Autonomous aerial vehicles; UAV; trajectory design; channel selection; MADQN; meta reinforcement learning;
D O I
10.23919/JCC.2022.04.013
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Unmanned aerial vehicle (UAV)-assisted communications have been considered as a solution of aerial networking in future wireless networks due to its low-cost, high-mobility, and swift features. This paper considers a UAV-assisted downlink transmission, where UAVs are deployed as aerial base stations to serve ground users. To maximize the average transmission rate among the ground users, this paper formulates a joint optimization problem of UAV trajectory design and channel selection, which is NP-hard and non-convex. To solve the problem, we propose a multi-agent deep Q-network (MADQN) scheme. Specifically, the agents that the UAVs act as perform actions from their observations distributively and share the same reward. To tackle the tasks where the experience is insufficient, we propose a multi-agent meta reinforcement learning algorithm to fast adapt to the new tasks. By pretraining the tasks with similar distribution, the learning model can acquire general knowledge. Simulation results have indicated the MADQN scheme can achieve higher throughput than fixed allocation. Furthermore, our proposed multiagent meta reinforcement learning algorithm learns the new tasks much faster compared with the MADQN scheme.
引用
收藏
页码:166 / 176
页数:11
相关论文
共 50 条
  • [1] Multi-Agent Few-Shot Meta Reinforcement Learning for Trajectory Design and Channel Selection in UAV-Assisted Networks
    Shiyang Zhou
    Yufan Cheng
    Xia Lei
    Huanhuan Duan
    ChinaCommunications, 2022, 19 (04) : 166 - 176
  • [2] Multi-Agent Low-Bias Reinforcement Learning for Resource Allocation in UAV-Assisted Networks
    Zhou, Shiyang
    Cheng, Yufan
    Lei, Xia
    2022 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2022, : 1011 - 1016
  • [3] UAV-Assisted Fair Communication for Mobile Networks: A Multi-Agent Deep Reinforcement Learning Approach
    Zhou, Yi
    Jin, Zhanqi
    Shi, Huaguang
    Wang, Zhangyun
    Lu, Ning
    Liu, Fuqiang
    REMOTE SENSING, 2022, 14 (22)
  • [4] Multi-Agent Reinforcement Learning Based Resource Management in MEC- and UAV-Assisted Vehicular Networks
    Peng, Haixia
    Shen, Xuemin
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (01) : 131 - 141
  • [5] Multi-Agent Deep Reinforcement Learning for Trajectory Design and Power Allocation in Multi-UAV Networks
    Zhao, Nan
    Liu, Zehua
    Cheng, Yiqiang
    IEEE ACCESS, 2020, 8 : 139670 - 139679
  • [6] Few-Shot Multi-Agent Perception
    Fan, Chenyou
    Hu, Junjie
    Huang, Jianwei
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 1712 - 1720
  • [7] Reinforcement Learning for Trajectory Design in Cache-enabled UAV-assisted Cellular Networks
    Xu, Hu
    Ji, Jiequ
    Zhu, Kun
    Wang, Ran
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 2238 - 2243
  • [8] GAN-powered heterogeneous multi-agent reinforcement learning for UAV-assisted task
    Li, Yangyang
    Feng, Lei
    Yang, Yang
    Li, Wenjing
    AD HOC NETWORKS, 2024, 153
  • [9] Age of information minimization in UAV-assisted data harvesting networks by multi-agent deep reinforcement curriculum learning
    Seong, Mincheol
    Jo, Ohyun
    Shin, Kyungseop
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [10] Joint UAV trajectory and communication design with heterogeneous multi-agent reinforcement learning
    Xuanhan ZHOU
    Jun XIONG
    Haitao ZHAO
    Xiaoran LIU
    Baoquan REN
    Xiaochen ZHANG
    Jibo WEI
    Hao YIN
    Science China(Information Sciences), 2024, 67 (03) : 225 - 245