Dense Multiagent Reinforcement Learning Aided Multi-UAV Information Coverage for Vehicular Networks

被引:5
|
作者
Fu, Hang [1 ,2 ]
Wang, Jingjing [1 ,2 ]
Chen, Jianrui [1 ,3 ]
Ren, Pengfei [1 ]
Zhang, Zheng [1 ]
Zhao, Guodong [4 ]
机构
[1] Beihang Univ, Sch Cyber Sci & Technol, Beijing 100191, Peoples R China
[2] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[3] Peng Cheng Lab, Shenzhen 518000, Peoples R China
[4] Beihang Univ, Sch Aeronaut Sci & Engn, Beijing 100191, Peoples R China
来源
IEEE INTERNET OF THINGS JOURNAL | 2024年 / 11卷 / 12期
关键词
Heuristic algorithms; Autonomous aerial vehicles; Vehicle dynamics; Training; Internet of Things; Energy consumption; Decision making; Communication coverage; dense reinforcement learning; distributed multiunmanned aerial vehicle (UAV); multiagent reinforcement learning (MARL); vehicular networks; RESOURCE-ALLOCATION; COMMUNICATION; OPTIMIZATION; ALTITUDE; INTERNET;
D O I
10.1109/JIOT.2024.3367005
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid development of wireless communication networks, UAVs serving as base stations are increasingly being applied in various scenarios which not only include edge computation and task offloading, but also involve emergency communication, vehicular network enhancement, etc. In order to enhance the utility of UAV base stations' allocation and deployment, a series of algorithms have been proposed, utilizing heuristic methods, learning-based algorithms, or optimization approaches. However, it is intractable for current algorithms to handle the exponential computation increment with UAV base stations increasing, and complicated application scenarios with high dynamic demands. To solve the above issues, we formulate a decision problem with a long sequence to optimize the deployment of multi-UAV base stations for maximizing vehicular networks' communication coverage ratio, which needs to be subject to co-constraints consisting of moving velocity, energy consumption, and communication coverage radius. To solve this optimization problem, we creatively propose an algorithm named dense multiagent reinforcement learning (DMARL), which is under the dual-layer nested decision-making framework, centralized training with decentralized deployment, and accelerates training by only collecting critical states into the dense sampling buffer. To prove our proposed algorithm's effectiveness and generalization ability, we conduct experimental simulations in scenarios with different scales. Corresponding results have been provided to verify our algorithm's superiority in training efficiency and performance metrics, including coverage ratio and energy consumption, compared with other algorithms.
引用
收藏
页码:21274 / 21286
页数:13
相关论文
共 50 条
  • [21] Resource Allocation for Multi-UAV Aided IoT NOMA Uplink Transmission Systems
    Duan, Ruiyang
    Wang, Jingjing
    Jiang, Chunxiao
    Yao, Haipeng
    Ren, Yong
    Qian, Yi
    IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (04) : 7025 - 7037
  • [22] AoI-Aware Joint Resource Allocation in Multi-UAV Aided Multi-Access Edge Computing Systems
    Shen, Shuai
    Yang, Halvin
    Yang, Kun
    Wang, Kezhi
    Zhang, Guopeng
    IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2024, 11 (03): : 2596 - 2609
  • [23] Distributed Energy-Efficient Multi-UAV Navigation for Long-Term Communication Coverage by Deep Reinforcement Learning
    Liu, Chi Harold
    Ma, Xiaoxin
    Gao, Xudong
    Tang, Jian
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2020, 19 (06) : 1274 - 1285
  • [24] Multi-Agent Reinforcement Learning in NOMA-Aided UAV Networks for Cellular Offloading
    Zhong, Ruikang
    Liu, Xiao
    Liu, Yuanwei
    Chen, Yue
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (03) : 1498 - 1512
  • [25] Cooperative Multiagent Deep Reinforcement Learning for Reliable Surveillance via Autonomous Multi-UAV Control
    Yun, Won Joon
    Park, Soohyun
    Kim, Joongheon
    Shin, MyungJae
    Jung, Soyi
    Mohaisen, David A.
    Kim, Jae-Hyun
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (10) : 7086 - 7096
  • [26] Resource Allocation and Trajectory Design in UAV-Aided Cellular Networks Based on Multiagent Reinforcement Learning
    Yin, Sixing
    Yu, F. Richard
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (04) : 2933 - 2943
  • [27] Multi-Agent Deep Reinforcement Learning for Trajectory Design and Power Allocation in Multi-UAV Networks
    Zhao, Nan
    Liu, Zehua
    Cheng, Yiqiang
    IEEE ACCESS, 2020, 8 : 139670 - 139679
  • [28] QoE-Driven Adaptive Deployment Strategy of Multi-UAV Networks Based on Hybrid Deep Reinforcement Learning
    Zhou, Yi
    Ma, Xiaoyong
    Hu, Shuting
    Zhou, Danyang
    Cheng, Nan
    Lu, Ning
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (08) : 5868 - 5881
  • [29] Integrating LEO Satellites and Multi-UAV Reinforcement Learning for Hybrid FSO/RF Non-Terrestrial Networks
    Lee, Ju-Hyung
    Park, Jihong
    Bennis, Mehdi
    Ko, Young-Chai
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (03) : 3647 - 3662
  • [30] Load Balance and Trajectory Design in Multi-UAV Aided Large-Scale Wireless Rechargeable Networks
    Wu, Pengfei
    Xiao, Fu
    Huang, Haiping
    Wang, Ruchuan
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (11) : 13756 - 13767