Integrated Localization and Tracking for AUV With Model Uncertainties via Scalable Sampling-Based Reinforcement Learning Approach

被引:34
作者
Yan, Jing [1 ]
Li, Xin [1 ]
Yang, Xian [2 ]
Luo, Xiaoyuan [1 ]
Hua, Changchun [1 ]
Guan, Xinping [3 ]
机构
[1] Yanshan Univ, Dept Automat, Qinhuangdao 066004, Hebei, Peoples R China
[2] Yanshan Univ, Inst Informat Sci & Engn, Qinhuangdao 066004, Hebei, Peoples R China
[3] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 11期
关键词
Location awareness; Uncertainty; Target tracking; Clocks; Computational modeling; Synchronization; Global Positioning System; Autonomous underwater vehicle (AUV); localization; reinforcement learning (RL); tracking; uncertain system; AUTONOMOUS UNDERWATER VEHICLE; TRAJECTORY TRACKING; JOINT LOCALIZATION; SENSOR NETWORKS; OPTIMIZATION;
D O I
10.1109/TSMC.2021.3129534
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies the joint localization and tracking issue for the autonomous underwater vehicle (AUV), with the constraints of asynchronous time clock in cyberchannels and model uncertainty in physical channels. More specifically, we develop a reinforcement learning (RL)-based asynchronous localization algorithm to localize the position of AUV, where the time clock of AUV is not required to be well synchronized with the real time. Based on the estimated position, a scalable sampling strategy called multivariate probabilistic collocation method with orthogonal fractional factorial design (M-PCM-OFFD) is employed to evaluate the time-varying uncertain model parameters of AUV. After that, an RL-based tracking controller is designed to drive AUV to the desired target point. Besides that, the performance analyses for the integration solution are also presented. Of note, the advantages of our solution are highlighted as: 1) the RL-based localization algorithm can avoid local optimal in traditional least-square methods; 2) the M-PCM-OFFD-based sampling strategy can address the model uncertainty and reduce the computational cost; and 3) the integration design of localization and tracking can reduce the communication energy consumption. Finally, simulation and experiment demonstrate that the proposed localization algorithm can effectively eliminate the impact of asynchronous clock, and more importantly, the integration of M-PCM-OFFD in the RL-based tracking controller can find accurate optimization solutions with limited computational costs.
引用
收藏
页码:6952 / 6967
页数:16
相关论文
共 50 条
  • [21] IoT Sensor Selection for Target Localization: A Reinforcement Learning based Approach
    Shurrab, Mohammed
    Singh, Shakti
    Mizouni, Rabeb
    Otrok, Hadi
    AD HOC NETWORKS, 2022, 134
  • [22] Local integrated energy system operational optimization considering multi-type uncertainties: A reinforcement learning approach based on improved TD3 algorithm
    Qiu, Yue
    Zhou, Suyang
    Xia, Dong
    Gu, Wei
    Sun, Kaiyu
    Han, Gaoyan
    Zhang, Kang
    Lv, Hongkun
    IET RENEWABLE POWER GENERATION, 2023, 17 (09) : 2236 - 2256
  • [23] Autonomous Underwater Vehicle (AUV) Motion Design: Integrated Path Planning and Trajectory Tracking Based on Model Predictive Control (MPC)
    Deng, Si-Yi
    Hao, Li-Ying
    Shen, Chao
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2024, 12 (09)
  • [24] Sliding mode heading control for AUV based on continuous hybrid model-free and model-based reinforcement learning
    Wang, Dianrui
    Shen, Yue
    Wan, Junhe
    Sha, Qixin
    Li, Guangliang
    Chen, Guanzhong
    He, Bo
    APPLIED OCEAN RESEARCH, 2022, 118
  • [25] Resilient Optimal Defensive Strategy of TSK Fuzzy-Model-Based Microgrids' System via a Novel Reinforcement Learning Approach
    Zhang, Huifeng
    Yue, Dong
    Dou, Chunxia
    Xie, Xiangpeng
    Li, Kang
    Hancke, Gerhardus P.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (04) : 1921 - 1931
  • [26] Safe Reinforcement Learning for Model-Reference Trajectory Tracking of Uncertain Autonomous Vehicles With Model-Based Acceleration
    Hu, Yifan
    Fu, Junjie
    Wen, Guanghui
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (03): : 2332 - 2344
  • [27] Reinforcement-learning-based fixed-time attitude consensus control for multiple spacecraft systems with model uncertainties
    Chen, Run-Ze
    Li, Yuan-Xin
    Ahn, Choon Ki
    AEROSPACE SCIENCE AND TECHNOLOGY, 2023, 132
  • [28] Consideration of uncertainties in a dynamic modeling system integrated with a deep learning based forecasting approach
    Biswas, Sumana
    Chakrabortty, Ripon K.
    Turan, Hasan Hueseyin
    Elsawah, Sondoss
    CIRP JOURNAL OF MANUFACTURING SCIENCE AND TECHNOLOGY, 2023, 44 : 27 - 44
  • [29] Integrated approach to AUV docking based on nonlinear offset-free model predictive control
    Shi, Kai
    Wang, Xiaohui
    Xu, Huixi
    Chen, Zhong
    Zhao, Hongyin
    MEASUREMENT & CONTROL, 2023, 56 (3-4) : 733 - 750
  • [30] Reinforcement Learning-Based Approach to Robot Path Tracking in Nonlinear Dynamic Environments
    Chen, Wei
    Zhou, Zebin
    INTERNATIONAL JOURNAL OF HUMANOID ROBOTICS, 2024, 21 (04)