Integrated Localization and Tracking for AUV With Model Uncertainties via Scalable Sampling-Based Reinforcement Learning Approach

被引:34
作者
Yan, Jing [1 ]
Li, Xin [1 ]
Yang, Xian [2 ]
Luo, Xiaoyuan [1 ]
Hua, Changchun [1 ]
Guan, Xinping [3 ]
机构
[1] Yanshan Univ, Dept Automat, Qinhuangdao 066004, Hebei, Peoples R China
[2] Yanshan Univ, Inst Informat Sci & Engn, Qinhuangdao 066004, Hebei, Peoples R China
[3] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 11期
关键词
Location awareness; Uncertainty; Target tracking; Clocks; Computational modeling; Synchronization; Global Positioning System; Autonomous underwater vehicle (AUV); localization; reinforcement learning (RL); tracking; uncertain system; AUTONOMOUS UNDERWATER VEHICLE; TRAJECTORY TRACKING; JOINT LOCALIZATION; SENSOR NETWORKS; OPTIMIZATION;
D O I
10.1109/TSMC.2021.3129534
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies the joint localization and tracking issue for the autonomous underwater vehicle (AUV), with the constraints of asynchronous time clock in cyberchannels and model uncertainty in physical channels. More specifically, we develop a reinforcement learning (RL)-based asynchronous localization algorithm to localize the position of AUV, where the time clock of AUV is not required to be well synchronized with the real time. Based on the estimated position, a scalable sampling strategy called multivariate probabilistic collocation method with orthogonal fractional factorial design (M-PCM-OFFD) is employed to evaluate the time-varying uncertain model parameters of AUV. After that, an RL-based tracking controller is designed to drive AUV to the desired target point. Besides that, the performance analyses for the integration solution are also presented. Of note, the advantages of our solution are highlighted as: 1) the RL-based localization algorithm can avoid local optimal in traditional least-square methods; 2) the M-PCM-OFFD-based sampling strategy can address the model uncertainty and reduce the computational cost; and 3) the integration design of localization and tracking can reduce the communication energy consumption. Finally, simulation and experiment demonstrate that the proposed localization algorithm can effectively eliminate the impact of asynchronous clock, and more importantly, the integration of M-PCM-OFFD in the RL-based tracking controller can find accurate optimization solutions with limited computational costs.
引用
收藏
页码:6952 / 6967
页数:16
相关论文
共 50 条
  • [31] Neural Network Model-Based Reinforcement Learning Control for AUV 3-D Path Following
    Ma, Dongfang
    Chen, Xi
    Ma, Weihao
    Zheng, Huarong
    Qu, Fengzhong
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 893 - 904
  • [32] Exploring the variance contributions of correlated model parameters: A sampling-based approach and its application in traffic simulation models
    Ge, Qiao
    Menendez, Monica
    APPLIED MATHEMATICAL MODELLING, 2021, 97 : 438 - 462
  • [33] Model-Based Reinforcement Learning for Infinite-Horizon Approximate Optimal Tracking
    Kamalapurkar, Rushikesh
    Andrews, Lindsey
    Walters, Patrick
    Dixon, Warren E.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) : 753 - 758
  • [34] Reinforcement Learning Tracking Control for Robotic Manipulator With Kernel-Based Dynamic Model
    Hu, Yazhou
    Wang, Wenxue
    Liu, Hao
    Liu, Lianqing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (09) : 3570 - 3578
  • [35] Wind-Farm Power Tracking Via Preview-Based Robust Reinforcement Learning
    Dong, Hongyang
    Zhao, Xiaowei
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (03) : 1706 - 1715
  • [36] A Localization and Tracking Approach in NLOS Environment Based on Distance and Angle Probability Model
    Tian, Xin
    Wei, Guoliang
    Wang, Jianhua
    Zhang, Dianchen
    SENSORS, 2019, 19 (20)
  • [37] A UNIVERSAL TRAJECTORY TRACKING CONTROLLER FOR MOBILE ROBOTS VIA MODEL-FREE ONLINE REINFORCEMENT LEARNING
    Fahimi, Farbod
    Praneeth, Susheel
    CONTROL AND INTELLIGENT SYSTEMS, 2015, 43 (01) : 56 - 64
  • [38] Practical Probabilistic Model-Based Reinforcement Learning by Integrating Dropout Uncertainty and Trajectory Sampling
    Huang, Wenjun
    Cui, Yunduan
    Li, Huiyun
    Wu, Xinyu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [39] INTEGRATED ACTIVE-REACTIVE POWER DISPATCHING MODEL BASED ON DEEP REINFORCEMENT LEARNING
    Lin, Zizhao
    Liu, Xuefei
    Tian, Qidong
    UNIVERSITY POLITEHNICA OF BUCHAREST SCIENTIFIC BULLETIN SERIES C-ELECTRICAL ENGINEERING AND COMPUTER SCIENCE, 2022, 84 (01): : 297 - 316
  • [40] Optimal scheduling of island integrated energy systems considering multi-uncertainties and hydrothermal simultaneous transmission: A deep reinforcement learning approach
    Li, Yang
    Bu, Fanjin
    Li, Yuanzheng
    Long, Chao
    APPLIED ENERGY, 2023, 333