Integrated Localization and Tracking for AUV With Model Uncertainties via Scalable Sampling-Based Reinforcement Learning Approach

被引:34
作者
Yan, Jing [1 ]
Li, Xin [1 ]
Yang, Xian [2 ]
Luo, Xiaoyuan [1 ]
Hua, Changchun [1 ]
Guan, Xinping [3 ]
机构
[1] Yanshan Univ, Dept Automat, Qinhuangdao 066004, Hebei, Peoples R China
[2] Yanshan Univ, Inst Informat Sci & Engn, Qinhuangdao 066004, Hebei, Peoples R China
[3] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2022年 / 52卷 / 11期
关键词
Location awareness; Uncertainty; Target tracking; Clocks; Computational modeling; Synchronization; Global Positioning System; Autonomous underwater vehicle (AUV); localization; reinforcement learning (RL); tracking; uncertain system; AUTONOMOUS UNDERWATER VEHICLE; TRAJECTORY TRACKING; JOINT LOCALIZATION; SENSOR NETWORKS; OPTIMIZATION;
D O I
10.1109/TSMC.2021.3129534
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This article studies the joint localization and tracking issue for the autonomous underwater vehicle (AUV), with the constraints of asynchronous time clock in cyberchannels and model uncertainty in physical channels. More specifically, we develop a reinforcement learning (RL)-based asynchronous localization algorithm to localize the position of AUV, where the time clock of AUV is not required to be well synchronized with the real time. Based on the estimated position, a scalable sampling strategy called multivariate probabilistic collocation method with orthogonal fractional factorial design (M-PCM-OFFD) is employed to evaluate the time-varying uncertain model parameters of AUV. After that, an RL-based tracking controller is designed to drive AUV to the desired target point. Besides that, the performance analyses for the integration solution are also presented. Of note, the advantages of our solution are highlighted as: 1) the RL-based localization algorithm can avoid local optimal in traditional least-square methods; 2) the M-PCM-OFFD-based sampling strategy can address the model uncertainty and reduce the computational cost; and 3) the integration design of localization and tracking can reduce the communication energy consumption. Finally, simulation and experiment demonstrate that the proposed localization algorithm can effectively eliminate the impact of asynchronous clock, and more importantly, the integration of M-PCM-OFFD in the RL-based tracking controller can find accurate optimization solutions with limited computational costs.
引用
收藏
页码:6952 / 6967
页数:16
相关论文
共 50 条
  • [41] Model-Based Chance-Constrained Reinforcement Learning via Separated Proportional-Integral Lagrangian
    Peng, Baiyu
    Duan, Jingliang
    Chen, Jianyu
    Li, Shengbo Eben
    Xie, Genjin
    Zhang, Congsheng
    Guan, Yang
    Mu, Yao
    Sun, Enxin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (01) : 466 - 478
  • [42] Model-Based Offline Reinforcement Learning for AUV Path-Following Under Unknown Ocean Currents with Limited Data
    Li, Xinmao
    Geng, Lingbo
    Liu, Kaizhou
    Zhao, Yifeng
    DRONES, 2025, 9 (03)
  • [43] Fast Human-in-the-Loop Control for HVAC Systems via Meta-Learning and Model-Based Offline Reinforcement Learning
    Chen, Liangliang
    Meng, Fei
    Zhang, Ying
    IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING, 2023, 8 (03): : 504 - 521
  • [44] Active reinforcement learning based approach for localization of target ROI (region of interest) in cervical cell images
    Rishi Khajuria
    Abid Sarwar
    Multimedia Tools and Applications, 2025, 84 (17) : 18467 - 18479
  • [45] Simulating human walking: a model-based reinforcement learning approach with musculoskeletal modeling
    Su, Binbin
    Gutierrez-Farewik, Elena M.
    FRONTIERS IN NEUROROBOTICS, 2023, 17
  • [46] Olfactory-Based Navigation via Model-Based Reinforcement Learning and Fuzzy Inference Methods
    Wang, Lingxiao
    Pang, Shuo
    Li, Jinlong
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2021, 29 (10) : 3014 - 3027
  • [47] Reinforcement learning approach for adapting complex Agent-Based model of evacuation to fast Linear Model
    Le, Van-Minh
    Vinh, Ho Tuong
    Zucker, Jean-Daniel
    2017 SEVENTH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST2017), 2017, : 369 - 375
  • [48] Building HVAC Scheduling Using Reinforcement Learning via Neural Network Based Model Approximation
    Zhang, Chi
    Kuppannagari, Sanmukh R.
    Kannan, Rajgopal
    Prasanna, Viktor K.
    BUILDSYS'19: PROCEEDINGS OF THE 6TH ACM INTERNATIONAL CONFERENCE ON SYSTEMS FOR ENERGY-EFFICIENT BUILDINGS, CITIES, AND TRANSPORTATION, 2019, : 287 - 296
  • [49] Eye in the Sky: Energy Efficient Model-Based Reinforcement Learning Aided Target Tracking Using UAVs
    Xia, Yi
    Zhang, Zekai
    Xu, Jingzehua
    Ren, Pengfei
    Wang, Jingjing
    Han, Zhu
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2024, 73 (12) : 19464 - 19479
  • [50] Deep Reinforcement Learning Based Acceleration Approach for Day-Ahead Optimal Dispatch of Integrated Energy Systems
    Lu, Yudong
    Yang, Miao
    Jia, Wenhao
    He, Xinran
    Fang, Yunhui
    Ding, Tao
    2023 5TH ASIA ENERGY AND ELECTRICAL ENGINEERING SYMPOSIUM, AEEES, 2023, : 1783 - 1789