AUV-Aided Localization for Internet of Underwater Things: A Reinforcement-Learning-Based Method

被引:69
作者
Yan, Jing [1 ]
Gong, Yadi [1 ]
Chen, Cailian [2 ]
Luo, Xiaoyuan [1 ]
Guan, Xinping [2 ]
机构
[1] Yanshan Univ, Dept Elect Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Shanghai Jiao Tong Univ, Dept Automat, Shanghai 200240, Peoples R China
来源
IEEE INTERNET OF THINGS JOURNAL | 2020年 / 7卷 / 10期
关键词
Clocks; Navigation; Internet of Things; Protocols; Unmanned aerial vehicles; Optimization; Acoustic communication; Internet of Underwater Things (IoUT); localization; reinforcement learning (RL); ASYNCHRONOUS LOCALIZATION; DISTRIBUTED LOCALIZATION; TARGET LOCALIZATION; JOINT LOCALIZATION; SENSOR NETWORKS; SYNCHRONIZATION; TRACKING;
D O I
10.1109/JIOT.2020.2993012
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Localization is a critical issue for many location-based applications in the Internet of Underwater Things (IoUT). Nevertheless, the asynchronous time clock, stratification effect, and mobility properties of the underwater environment make it much more challenging to solve the localization issue. This article is concerned with an autonomous underwater vehicle (AUV)-aided localization issue for IoUT. We first provide a hybrid network architecture that includes surface buoys, AUVs, and active and passive sensor nodes. On the basis of this architecture, an asynchronous localization protocol is designed, through which the localization problem is provided to minimize the sum of all measurement errors. In order to make this problem tractable, a reinforcement-learning (RL)-based localization algorithm is developed to estimate the locations of AUVs, and active and passive sensor nodes, where an online value iteration procedure is performed to seek the optimization locations. It is worth mentioning that the proposed localization algorithm adopts two neural networks to approximate the increment policy and value function, and more importantly, it is much preferable for the nonsmooth and nonconvex underwater localization problem due to its insensitivity to the local optimal. Performance analyses for the RL-based localization algorithm are also provided. Finally, simulation and experimental results reveal that the localization performance in this article can be significantly improved as compared with the other works.
引用
收藏
页码:9728 / 9746
页数:19
相关论文
共 57 条
[1]  
Akyildiz I. F., 2004, ACM SIGBED Rev, V1, P3, DOI DOI 10.1145/1121776.1121779
[2]  
Aras MSM, 2012, MAJLESI J ELECT ENG, V6, P1
[3]   Square-root quadrature Kalman filtering [J].
Arasaratnam, Ienkaran ;
Haykin, Simon .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2008, 56 (06) :2589-2593
[4]   AN ANALYTICAL STUDY OF CHAOTIC STIRRING IN TIDAL AREAS [J].
BEERENS, SP ;
RIDDERINKHOF, H ;
ZIMMERMAN, JTF .
CHAOS SOLITONS & FRACTALS, 1994, 4 (06) :1011-1029
[5]   An overview of the internet of underwater things [J].
Carmen Domingo, Mari .
JOURNAL OF NETWORK AND COMPUTER APPLICATIONS, 2012, 35 (06) :1879-1890
[6]   On-Demand Asynchronous Localization for Underwater Sensor Networks [J].
Carroll, Patrick ;
Mahmood, Kaleel ;
Zhou, Shengli ;
Zhou, Hao ;
Xu, Xiaoka ;
Cui, Jun-Hong .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2014, 62 (13) :3337-3348
[7]   Underwater Localization with Time-Synchronization and Propagation Speed Uncertainties [J].
Diamant, Roee ;
Lampe, Lutz .
IEEE TRANSACTIONS ON MOBILE COMPUTING, 2013, 12 (07) :1257-1269
[8]   Adaptive Partial Reinforcement Learning Neural Network-Based Tracking Control for Wheeled Mobile Robotic Systems [J].
Ding, Liang ;
Li, Shu ;
Gao, Haibo ;
Chen, Chao ;
Deng, Zongquan .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (07) :2512-2523
[9]  
Dunn CW, 2012, AMERICAN CULTURE IN PERIL, P1
[10]   Supervised fuzzy reinforcement learning for robot navigation [J].
Fathinezhad, Fatemeh ;
Derhami, Vali ;
Rezaeian, Mehdi .
APPLIED SOFT COMPUTING, 2016, 40 :33-41