AUV Obstacle Avoidance and Path Planning Based on Artificial Potential Field and Improved Reinforcement Learning

被引:0
作者
Pan, Yunwei [1 ]
Li, Min [1 ]
Zeng, Xiangguang [1 ]
Huang, Ao [1 ]
Zhang, Jiaheng [1 ]
Ren, Wenzhe [1 ]
Peng, Bei [2 ]
机构
[1] School of Mechanical Engineering, Southwest Jiaotong University, Sichuan, Chengdu
[2] School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Sichuan, Chengdu
来源
Binggong Xuebao/Acta Armamentarii | 2025年 / 46卷 / 04期
关键词
artificial potential field; autonomous underwater vehicle; obstacle avoidance; path planning; reinforcement learning;
D O I
10.12382/bgxb.2024.0300
中图分类号
学科分类号
摘要
Autonomous underwater vehicle (AUV),as one of the important underwater detection tools,are widely used in various marine military operations. Most of the existing research on AUV obstacle avoidance and path planning focuses on grid maps,and rarely considers the real maneuverability of AUVs under water. In order to solve this problem,an improved proximal policy optimization based on positive-experience retraining (PR-PPO) algorithm and an AUV obstacle avoidance and path planning method based on artificial potential field are proposed. A dynamic artificial potential field is constructed by using the sensor in AUV model and the underwater environment in the simulation software. Based on the PR-PPO reinforcement learning algorithm,the mapping relationship between the AUV state and the action is established by interacting with the environment. Real-time obstacle avoidance and path planning can be realized without dynamic model and map information. The results show that,compared with the traditional D3QN and PPO algorithms,the proposed algorithm can not only ensure the success rate of the task,but also shorten the model training time and improve the convergence effect. © 2025 China Ordnance Industry Corporation. All rights reserved.
引用
收藏
相关论文
共 21 条
[1]  
WANG S J, KANG F J, HAN H., Research on control and decision-making of submarine and intelligent UUV cooperative system, Acta Armamentarii, 38, 2, pp. 335-344, (2017)
[2]  
DING W J, ZHANG G Z, LIU H M, Et al., Tracking control of underactuated autonomous underwater vehicle formation under current disturbance and communication delay [ J ], Acta Armamentarii, 45, 1, pp. 184-196, (2024)
[3]  
WANG C, CHENG C S, YANG D Y, Et al., Path planning in localization uncertaining environment based on Dijkstra method [J], Frontiers in Neurorobotics, 16, (2022)
[4]  
LIU Y, HUANG P F, ZHANG F, Et al., Distributed formation control using artificial potentials and neural network for constrained multiagent systems [ J], IEEE Transactions on Control Systems Technology, 28, 2, pp. 697-704, (2020)
[5]  
WANG W, ZUO L, XU X., A learning-based multi-RRT approach for robot path planning in narrow passages [ J ], Journal of Intelligent and Robotic Systems:Theory and Applications, 90, 1, pp. 81-100, (2018)
[6]  
TAO C L, ZHANG Y R, LI S Q, Et al., Mechanism of interior ballistic peak phenomenon of guns and its effects[ J], Journal of Applied Mechanics, 77, 5, (2010)
[7]  
ROBERGE V, TARBOUCHI M, LABONTE G., Comparison of parallel genetic algorithm and particle swarm optimization for real-time UAV path planning [ J], IEEE Transactions on Industrial Informatics, 9, 1, pp. 132-141, (2012)
[8]  
CHEN M Z, ZHU D Q., Optimal time-consuming path planning for autonomous underwater vehicles based on a dynamic neural network model in ocean current environments [ J], IEEE Transactions on Vehicular Technology, 69, 12, pp. 14401-14412, (2020)
[9]  
YANG J, WU J P, LIU J, Et al., A semi-supervised learning method for intelligent decision making of submarine maneuver defense, Acta Armamentarii, 45, 10, pp. 3474-3487, (2024)
[10]  
HE Z C, DONG L, SUN C Y, Et al., Asynchronous multithreading reinforcement-learning-based path planning and tracking for unmanned underwater vehicle[J], IEEE Transactions on Systems, Man,and Cybernetics:Systems, 52, 5, pp. 2757-2769, (2022)