A reinforcement learning-based evolutionary algorithm for the unmanned aerial vehicles maritime search and rescue path planning problem considering multiple rescue centers

被引:2
|
作者
Zhan, Haowen [1 ]
Zhang, Yue [2 ]
Huang, Jingbo [1 ]
Song, Yanjie [3 ]
Xing, Lining [4 ]
Wu, Jie [5 ]
Gao, Zengyun [6 ]
机构
[1] Natl Univ Def Technol, Coll Syst Engn, Changsha 410073, Peoples R China
[2] Beihang Univ, Sch Reliabil & Syst Engn, Beijing 100191, Peoples R China
[3] Wuyi Intelligent Mfg Inst Ind Technol, Jinhua 321017, Peoples R China
[4] Xidian Univ, Key Lab Collaborat Intelligence Syst, Xian 710071, Peoples R China
[5] Nanjing Univ, Sch Geog & Ocean Sci, Nanjing 210023, Peoples R China
[6] China Maritime Serv Ctr, Beijing 100029, Peoples R China
基金
中国国家自然科学基金;
关键词
UAV; Maritime search and rescue; Path planning; Reinforcement learning; Evolutionary algorithm; Genetic;
D O I
10.1007/s12293-024-00420-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the realm of maritime emergencies, unmanned aerial vehicles (UAVs) play a crucial role in enhancing search and rescue (SAR) operations. They help in efficiently rescuing distressed crews, strengthening maritime surveillance, and maintaining national security due to their cost-effectiveness, versatility, and effectiveness. However, the vast expanse of sea territories and the rapid changes in maritime conditions make a single SAR center insufficient for handling complex emergencies. Thus, it is vital to develop strategies for quickly deploying UAV resources from multiple SAR centers for area reconnaissance and supporting maritime rescue operations. This study introduces a graph-structured planning model for the maritime SAR path planning problem, considering multiple rescue centers (MSARPPP-MRC). It incorporates workload distribution among SAR centers and UAV operational constraints. We propose a reinforcement learning-based genetic algorithm (GA-RL) to tackle the MSARPPP-MRC problem. GA-RL uses heuristic rules to initialize the population and employs the Q-learning method to manage the progeny during each generation, including their retention, storage, or disposal. When the elite repository's capacity is reached, a decision is made on the utilization of these members to refresh the population. Additionally, adaptive crossover and perturbation strategies are applied to develop a more effective SAR scheme. Extensive testing proves that GA-RL surpasses other algorithms in optimization efficacy and efficiency, highlighting the benefits of reinforcement learning in population management.
引用
收藏
页码:373 / 386
页数:14
相关论文
共 50 条
  • [31] Learning-based Preferential Surveillance Algorithm for Persistent Surveillance by Unmanned Aerial Vehicles
    Ramasamy, Manickam
    Ghose, Debasish
    2016 INTERNATIONAL CONFERENCE ON UNMANNED AIRCRAFT SYSTEMS (ICUAS), 2016, : 1032 - 1040
  • [32] An immune plasma algorithm with Q-learning based pandemic management for path planning of unmanned aerial vehicles
    Aslan, Selcuk
    Demirci, Sercan
    EGYPTIAN INFORMATICS JOURNAL, 2024, 26
  • [33] Multi-Objective Cooperated Path Planning of Multiple Unmanned Aerial Vehicles Based on Revisit Time
    Haghighi, Hassan
    Asadi, Davood
    Delahaye, Daniel
    JOURNAL OF AEROSPACE INFORMATION SYSTEMS, 2021, : 919 - 932
  • [34] Autonomous Obstacle Avoidance Algorithm for Unmanned Aerial Vehicles Based on Deep Reinforcement Learning
    Gao, Yuan
    Ren, Ling
    Shi, Tianwei
    Xu, Teng
    Ding, Jianbang
    ENGINEERING LETTERS, 2024, 32 (03) : 650 - 660
  • [35] Multi-UAV Coverage Path Planning Based on Hexagonal Grid Decomposition in Maritime Search and Rescue
    Cho, Sung-Won
    Park, Jin-Hyoung
    Park, Hyun-Ji
    Kim, Seongmin
    MATHEMATICS, 2022, 10 (01)
  • [36] Research on Subareas Division Problem of Maritime Search and Rescue Based on Improved Monte Carlo Algorithm
    Jie, Li
    Yao, Sun
    Zhang, Zhenxing
    2013 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION (ICMA), 2013, : 1464 - 1468
  • [37] Improved Q-Learning Algorithm Based on Flower Pollination Algorithm and Tabulation Method for Unmanned Aerial Vehicle Path Planning
    Bo, Lan
    Zhang, Tiezhu
    Zhang, Hongxin
    Yang, Jian
    Zhang, Zhen
    Zhang, Caihong
    Liu, Mingjie
    IEEE ACCESS, 2024, 12 : 104429 - 104444
  • [38] Reinforcement learning-based multi-strategy cuckoo search algorithm for 3D UAV path planning
    Yu, Xiaobing
    Luo, Wenguan
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 223
  • [39] Unmanned Aerial Vehicle Path Planning Algorithm Based on Deep Reinforcement Learning in Large-Scale and Dynamic Environments
    Xie, Ronglei
    Meng, Zhijun
    Wang, Lifeng
    Li, Haochen
    Wang, Kaipeng
    Wu, Zhe
    IEEE Access, 2021, 9 : 24884 - 24900
  • [40] A Long-Term Target Search Method for Unmanned Aerial Vehicles Based on Reinforcement Learning
    Wei, Dexing
    Zhang, Lun
    Yang, Mei
    Deng, Hanqiang
    Huang, Jian
    DRONES, 2024, 8 (10)