Dynamic Navigation and Area Assignment of Multiple USVs Based on Multi-Agent Deep Reinforcement Learning

被引:9
作者
Wen, Jiayi [1 ]
Liu, Shaoman [1 ]
Lin, Yejin [1 ]
机构
[1] Dalian Maritime Univ, Lab Intelligent Marine Vehicles DMU, Dalian 116026, Peoples R China
基金
中国国家自然科学基金;
关键词
USV; trajectory design; policy gradient; multi-agent deep reinforcement learning; multi-object optimization; SURFACE; TRACKING; VEHICLE;
D O I
10.3390/s22186942
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The unmanned surface vehicle (USV) has attracted more and more attention because of its basic ability to perform complex maritime tasks autonomously in constrained environments. However, the level of autonomy of one single USV is still limited, especially when deployed in a dynamic environment to perform multiple tasks simultaneously. Thus, a multi-USV cooperative approach can be adopted to obtain the desired success rate in the presence of multi-mission objectives. In this paper, we propose a cooperative navigating approach by enabling multiple USVs to automatically avoid dynamic obstacles and allocate target areas. To be specific, we propose a multi-agent deep reinforcement learning (MADRL) approach, i.e., a multi-agent deep deterministic policy gradient (MADDPG), to maximize the autonomy level by jointly optimizing the trajectory of USVs, as well as obstacle avoidance and coordination, which is a complex optimization problem usually solved separately. In contrast to other works, we combined dynamic navigation and area assignment to design a task management system based on the MADDPG learning framework. Finally, the experiments were carried out on the Gym platform to verify the effectiveness of the proposed method.
引用
收藏
页数:14
相关论文
共 39 条
  • [31] Mode Selection and Resource Allocation in Sliced Fog Radio Access Networks: A Reinforcement Learning Approach
    Xiang, Hongyu
    Peng, Mugen
    Sun, Yaohua
    Yan, Shi
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (04) : 4271 - 4284
  • [32] A Sampling-Based Bayesian Approach for Cooperative Multiagent Online Search With Resource Constraints
    Xiao, Hu
    Cui, Rongxin
    Xu, Demin
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2018, 48 (06) : 1773 - 1785
  • [33] Xie SR, 2014, 2014 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION (ICIA), P746, DOI 10.1109/ICInfA.2014.6932751
  • [34] Obstacle-avoiding path planning for multiple autonomous underwater vehicles with simultaneous arrival
    Yao Peng
    Qi ShengBo
    [J]. SCIENCE CHINA-TECHNOLOGICAL SCIENCES, 2019, 62 (01) : 121 - 132
  • [35] Deep Reinforcement Learning Based Resource Allocation for V2V Communications
    Ye, Hao
    Li, Geoffrey Ye
    Juang, Biing-Hwang Fred
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (04) : 3163 - 3173
  • [36] Shell space decomposition based path planning for AUVs operating in a variable environment
    Zeng, Zheng
    Lammas, Andrew
    Sammut, Karl
    He, Fangpo
    Tang, Youhong
    [J]. OCEAN ENGINEERING, 2014, 91 : 181 - 195
  • [37] Real-time trajectory planning for UCAV air-to-surface attack using inverse dynamics optimization method and receding horizon control
    Zhang Yu
    Chen Jing
    Shen Lincheng
    [J]. CHINESE JOURNAL OF AERONAUTICS, 2013, 26 (04) : 1038 - 1056
  • [38] Secure UAV Communication With Cooperative Jamming and Trajectory Control
    Zhong, Canhui
    Yao, Jianping
    Xu, Jie
    [J]. IEEE COMMUNICATIONS LETTERS, 2019, 23 (02) : 286 - 289
  • [39] UAV-Enabled Secure Communications: Joint Trajectory and Transmit Power Optimization
    Zhou, Xiaobo
    Wu, Qingqing
    Yan, Shihao
    Shu, Feng
    Li, Jun
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (04) : 4069 - 4073