Generative adversarial interactive imitation learning for path following of autonomous underwater vehicle

被引:10
|
作者
Jiang, Dong [1 ]
Huang, Jie [1 ]
Fang, Zheng [1 ]
Cheng, Chunxi [1 ]
Sha, Qixin [1 ]
He, Bo [1 ]
Li, Guangliang [1 ]
机构
[1] Ocean Univ China, Coll Elect Engn, Qingdao, Peoples R China
关键词
Deep reinforcement learning; Autonomous control; Autonomous underwater vehicle; Imitation learning; Interactive reinforcement learning; LEVEL CONTROL; PID CONTROL; DEEP;
D O I
10.1016/j.oceaneng.2022.111971
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
Autonomous underwater vehicle (AUV) is playing a more and more important role in marine scientific research and resource exploration due to its flexibility. Recently, deep reinforcement learning (DRL) has been used to improve the autonomy of AUV. However, it is very time-consuming and even unpractical to define efficient reward functions for DRL to learn control policies in various tasks. In this paper, we implemented the generative adversarial imitation learning (GAIL) algorithm learning from demonstrated trajectories and proposed GA2IL learning from demonstrations and additional human rewards for AUV path following. We evaluated GAIL and our GA2IL method in a straight line following task and a sinusoids curve following task on the Gazebo platform extended to simulated underwater environments with AUV simulator of our lab. Both methods were compared to PPO-a classic traditional deep reinforcement learning from a predefined reward function, and a well-tuned PID controller. In addition, to evaluate the generalization of GAIL and our GA2IL method, we tested the trained control policies of the previous two tasks via GAIL and GA2IL in a new complex comb scan following task and a different sinusoids curve following task respectively. Our simulation results show AUV path following with GA2IL and GAIL can obtain a performance at a similar level to PPO and PID controller in both tasks. Moreover, GA2IL can generalize as well as PPO, adapting better to complex and different tasks than traditional PID controller.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Path Following Model Predictive Control of a Coupled Autonomous Underwater Vehicle
    Jimoh, Isah A.
    Yue, Hong
    IFAC PAPERSONLINE, 2024, 58 (20): : 183 - 188
  • [22] A robust model predictive path following controller for an Autonomous Underwater Vehicle
    Rath, Biranchi Narayan
    Subudhi, Bidyadhar
    OCEAN ENGINEERING, 2022, 244
  • [23] A new method of path following control of AUV (autonomous underwater vehicle)
    Wang, Yintao
    Zheng, Meiyun
    Yan, Weisheng
    Xibei Gongye Daxue Xuebao/Journal of Northwestern Polytechnical University, 2009, 27 (04): : 517 - 521
  • [24] Path Planning for the Autonomous Underwater Vehicle
    Kirsanov, Andrey
    Anavatti, Sreenatha G.
    Ray, Tapabrata
    SWARM, EVOLUTIONARY, AND MEMETIC COMPUTING, PT II (SEMCCO 2013), 2013, 8298 : 476 - 486
  • [25] Randomized Adversarial Imitation Learning for Autonomous Driving
    Shin, MyungJae
    Kim, Joongheon
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4590 - 4596
  • [26] Robot Manipulation Learning Using Generative Adversarial Imitation Learning
    Jabri, Mohamed Khalil
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4893 - 4894
  • [27] A Survey of Imitation Learning Based on Generative Adversarial Nets
    Lin J.-H.
    Zhang Z.-Z.
    Jiang C.
    Hao J.-Y.
    Jisuanji Xuebao/Chinese Journal of Computers, 2020, 43 (02): : 326 - 351
  • [28] Modeling Car-Following Behaviors and Driving Styles with Generative Adversarial Imitation Learning
    Zhou, Yang
    Fu, Rui
    Wang, Chang
    Zhang, Ruibin
    SENSORS, 2020, 20 (18) : 1 - 20
  • [29] Ranking-Based Generative Adversarial Imitation Learning
    Shi, Zhipeng
    Zhang, Xuehe
    Fang, Yu
    Li, Changle
    Liu, Gangfeng
    Zhao, Jie
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (10): : 8967 - 8974
  • [30] Generative Adversarial Imitation Learning from Failed Experiences
    Zhu, Jiacheng
    Lin, Jiahao
    Wang, Meng
    Chen, Yingfeng
    Fan, Changjie
    Jiang, Chong
    Zhang, Zongzhang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13997 - 13998