Risk-aware deep reinforcement learning for mapless navigation of unmanned surface vehicles in uncertain and congested environments

被引:0
作者
Wu, Xiangyu [1 ]
Wei, Changyun [1 ]
Guan, Dawei [2 ]
Ji, Ze [3 ]
机构
[1] Hohai Univ, Coll Mech & Elect Engn, Changzhou 213200, Jiangsu, Peoples R China
[2] Hohai Univ, Coll Harbour Coastal & Offshore Engn, Nanjing 210098, Jiangsu, Peoples R China
[3] Cardiff Univ, Sch Engn, Cardiff CF24 3AA, Wales
关键词
Deep reinforcement learning; Unmanned surface vehicles; Collision avoidance; Sensor-level navigation;
D O I
10.1016/j.oceaneng.2025.120446
中图分类号
U6 [水路运输]; P75 [海洋工程];
学科分类号
0814 ; 081505 ; 0824 ; 082401 ;
摘要
This paper addresses the navigation problem of Unmanned Surface Vehicles (USVs) in uncertain and congested environments. While previous research has extensively explored USV navigation, most approaches assume that the environmental maps and obstacle locations are pre-known to the USVs. In this paper, we focus on a sensor-level navigation approach that utilizes real-time LiDAR data integrated with deep reinforcement learning (DRL) for decision-making. To tackle sparse reward challenges, we propose a potential-based reward-shaping (PRS) module to regulate navigation behavior, and this module helps to improve the training efficiency of the twin delayed deep deterministic policy gradient (TD3) algorithm. Moreover, we introduce a risk evaluation and correction (REC) module to mitigate potential risks. This module employs a risk evaluation network to enhance the agent's risk awareness and an action-level correction mechanism to avoid unsafe behavior. The proposed approach is validated through ablation studies and comparative experiments in OpenAI Gym-based environments and simulated island regions of Zhoushan. The results indicate that the proposed approach significantly improves training efficiency while maintaining consistency and robustness in unknown and congested marine environments.
引用
收藏
页数:14
相关论文
共 25 条
  • [1] Barrera C., Padron I., Luis F., Llinas O., Trends and challenges in unmanned surface vehicles (USV): From survey to shipping, Int. J. Mar. Navig. Saf. Sea Transp., 15, 1, pp. 135-142, (2021)
  • [2] Bhatt A., Palenicek D., Belousov B., Argus M., Amiranashvili A., Brox T., Peters J., CrossQ: Batch normalization in deep reinforcement learning for greater sample efficiency and simplicity, International Conference on Learning Representations, ICLR, (2024)
  • [3] Cui Y., Osaki S., Matsubara T., Autonomous boat driving system using sample-efficient model predictive control-based reinforcement learning approach, J. Field Robotics, 38, 3, pp. 331-354, (2021)
  • [4] Cui Y., Peng L., Li H., Filtered probabilistic model predictive control-based reinforcement learning for unmanned surface vehicles, IEEE Trans. Ind. Inform., 18, 10, pp. 6950-6961, (2022)
  • [5] Fan Y., Sun Z., Wang G., A novel intelligent collision avoidance algorithm based on deep reinforcement learning approach for USV, Ocean Eng., 287, (2023)
  • [6] Fossen T.I., Handbook of Marine Craft Hydrodynamics and Motion Control, (2011)
  • [7] Fujimoto S., Chang W.-D., Smith E.J., Gu S.S., Precup D., Meger D., For SALE: State-action representation learning for deep reinforcement learning, Proceedings of the 37th International Conference on Neural Information Processing Systems, 36, pp. 61573-61624, (2023)
  • [8] Fujimoto S., Hoof H., Meger D., Addressing function approximation error in actor-critic methods, International Conference on Machine Learning, pp. 1587-1596, (2018)
  • [9] Gonzalez-Garcia A., Castaneda H., Garrido L., USV path-following control based on deep reinforcement learning and adaptive control, Global Oceans 2020: Singapore–US Gulf Coast, pp. 1-7, (2020)
  • [10] Guan W., Wang K., Autonomous collision avoidance of unmanned surface vehicles based on improved A-star and dynamic window approach algorithms, IEEE Intell. Transp. Syst. Mag., 15, 3, pp. 36-50, (2023)