An Information-Assisted Deep Reinforcement Learning Path Planning Scheme for Dynamic and Unknown Underwater Environment

被引:6
作者
Xi, Meng [1 ]
Yang, Jiachen [1 ]
Wen, Jiabao [1 ]
Li, Zhengjian [1 ]
Lu, Wen [2 ]
Gao, Xinbo [2 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin 300072, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
基金
中国国家自然科学基金;
关键词
Heuristic algorithms; Path planning; Reinforcement learning; Robustness; Neural networks; Vehicle dynamics; Oceans; Autonomous underwater vehicle (AUV); dynamic environment; path planning; reinforcement learning; robustness; TRACKING CONTROL; VEHICLES; ALGORITHM; LEVEL; AUV;
D O I
10.1109/TNNLS.2023.3332172
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An autonomous underwater vehicle (AUV) has shown impressive potential and promising exploitation prospects in numerous marine missions. Among its various applications, the most essential prerequisite is path planning. Although considerable endeavors have been made, there are several limitations. A complete and realistic ocean simulation environment is critically needed. As most of the existing methods are based on mathematical models, they suffer from a large gap with reality. At the same time, the dynamic and unknown environment places high demands on robustness and generalization. In order to overcome these limitations, we propose an information-assisted reinforcement learning path planning scheme. First, it performs numerical modeling based on real ocean current observations to establish a complete simulation environment with the grid method, including 3-D terrain, dynamic currents, local information, and so on. Next, we propose an information compression (IC) scheme to trim the mutual information (MI) between reinforcement learning neural network layers to improve generalization. A proof based on information theory provides solid support for this. Moreover, for the dynamic characteristics of the marine environment, we elaborately design a confidence evaluator (CE), which evaluates the correlation between two adjacent frames of ocean currents to provide confidence for the action. The performance of our method has been evaluated and proven by numerical results, which demonstrate a fair sensitivity to ocean currents and high robustness and generalization to cope with the dynamic and unknown underwater environment.
引用
收藏
页码:842 / 853
页数:12
相关论文
共 42 条
  • [1] Information Dropout: Learning Optimal Representations Through Noisy Computation
    Achille, Alessandro
    Soatto, Stefano
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (12) : 2897 - 2905
  • [2] Multi-AUV Target Search Based on Bioinspired Neurodynamics Model in 3-D Underwater Environments
    Cao, Xiang
    Zhu, Daqi
    Yang, Simon X.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (11) : 2364 - 2374
  • [3] Optimal Time-Consuming Path Planning for Autonomous Underwater Vehicles Based on a Dynamic Neural Network Model in Ocean Current Environments
    Chen, Mingzhi
    Zhu, Daqi
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (12) : 14401 - 14412
  • [4] Adaptive Neural Network Control of AUVs With Control Input Nonlinearities Using Reinforcement Learning
    Cui, Rongxin
    Yang, Chenguang
    Li, Yang
    Sharma, Sanjay
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2017, 47 (06): : 1019 - 1029
  • [5] European Centre for Medium-Range Weather Forecasts, About us
  • [6] Fujimoto S, 2018, PR MACH LEARN RES, V80
  • [7] An AUV-Assisted Data Gathering Scheme Based on Clustering and Matrix Completion for Smart Ocean
    Huang, Mingfeng
    Zhang, Kuan
    Zeng, Zhiwen
    Wang, Tian
    Liu, Yuxin
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (10) : 9904 - 9918
  • [8] IRI/LDEO Climate Data Library, About us
  • [9] Attention-Based Meta-Reinforcement Learning for Tracking Control of AUV With Time-Varying Dynamics
    Jiang, Peng
    Song, Shiji
    Huang, Gao
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) : 6388 - 6401
  • [10] Hierarchical and Stable Multiagent Reinforcement Learning for Cooperative Navigation Control
    Jin, Yue
    Wei, Shuangqing
    Yuan, Jian
    Zhang, Xudong
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (01) : 90 - 103