Deep reinforcement learning for time-critical wilderness search and rescue using drones

被引:0
|
作者
Ewers, Jan-Hendrik [1 ]
Anderson, David [1 ]
Thomson, Douglas [1 ]
机构
[1] Univ Glasgow, Autonomous Syst & Connect, Glasgow City, Scotland
来源
FRONTIERS IN ROBOTICS AND AI | 2025年 / 11卷
基金
英国工程与自然科学研究理事会;
关键词
reinforcement learning; search planning; mission planning; autonomous systems; wilderness search and rescue; unmanned aerial vehicle; machine learning;
D O I
10.3389/frobt.2024.1527095
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Traditional search and rescue methods in wilderness areas can be time-consuming and have limited coverage. Drones offer a faster and more flexible solution, but optimizing their search paths is crucial for effective operations. This paper proposes a novel algorithm using deep reinforcement learning to create efficient search paths for drones in wilderness environments. Our approach leverages a priori data about the search area and the missing person in the form of a probability distribution map. This allows the policy to learn optimal flight paths that maximize the probability of finding the missing person quickly. Experimental results show that our method achieves a significant improvement in search times compared to traditional coverage planning and search planning algorithms by over 160 % , a difference that can mean life or death in real-world search operations Additionally, unlike previous work, our approach incorporates a continuous action space enabled by cubature, allowing for more nuanced flight patterns.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Small Target Detection for Search and Rescue Operations using Distributed Deep Learning and Synthetic Data Generation
    Yun, Kyongsik
    Luan Nguyen
    Tuan Nguyen
    Kim, Doyoung
    Eldin, Sarah
    Huyen, Alexander
    Lu, Thomas
    Chow, Edward
    PATTERN RECOGNITION AND TRACKING XXX, 2019, 10995
  • [32] Target Search Control of AUV in Underwater Environment With Deep Reinforcement Learning
    Cao, Xiang
    Sun, Changyin
    Yan, Mingzhong
    IEEE ACCESS, 2019, 7 : 96549 - 96559
  • [33] Deep Belief Network Using Reinforcement Learning and Its Applications to Time Series Forecasting
    Hirata, Takaomi
    Kuremoto, Takashi
    Obayashi, Masanao
    Mabu, Shingo
    Kobayashi, Kunikazu
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT III, 2016, 9949 : 30 - 37
  • [34] Forecasting Real Time Series Data using Deep Belief Net and Reinforcement Learning
    Hirata, Takaomi
    Kuremoto, Takashi
    Obayashi, Masanao
    Mabu, Shingo
    Kobayashi, Kunikazu
    JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2018, 4 (04): : 260 - 264
  • [35] Forecasting Real Time Series Data using Deep Belief Net and Reinforcement Learning
    Hirata, Takaomi
    Kuremoto, Takashi
    Obayashi, Masanao
    Mabu, Shingo
    Kobayashi, Kunikazu
    ICAROB 2017: PROCEEDINGS OF THE 2017 INTERNATIONAL CONFERENCE ON ARTIFICIAL LIFE AND ROBOTICS, 2017, : P658 - P661
  • [36] Quantitative analysis of EXAFS data sets using deep reinforcement learning
    Eun-Suk Jeong
    In-Hui Hwang
    Sang-Wook Han
    Scientific Reports, 15 (1)
  • [37] Aircraft collision avoidance modeling and optimization using deep reinforcement learning
    Park K.-W.
    Kim J.-H.
    Journal of Institute of Control, Robotics and Systems, 2021, 27 (09) : 652 - 659
  • [38] Mobile Service Robot Path Planning Using Deep Reinforcement Learning
    Kumaar, A. A. Nippun
    Kochuvila, Sreeja
    IEEE ACCESS, 2023, 11 : 100083 - 100096
  • [39] Advanced planning for autonomous vehicles using reinforcement learning and deep inverse reinforcement learning
    You, Changxi
    Lu, Jianbo
    Filev, Dimitar
    Tsiotras, Panagiotis
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 114 : 1 - 18
  • [40] PrefixRL: Optimization of Parallel Prefix Circuits using Deep Reinforcement Learning
    Roy, Rajarshi
    Raiman, Jonathan
    Kant, Neel
    Elkin, Ilyas
    Kirby, Robert
    Siu, Michael
    Oberman, Stuart
    Godil, Saad
    Catanzaro, Bryan
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 853 - 858