Risk-averse Ambulance Redeployment via Multi-armed Bandits

被引:0
|
作者
Sahin, Umitcan [1 ,2 ]
Yucesoy, Veysel [1 ]
Koc, Aykut [1 ]
Tekin, Cem [2 ]
机构
[1] Aselsan Arastirma Merkezi, Akilli Veri Analit Arastirma Program Mudurlugu, TR-06370 Ankara, Turkey
[2] Bilkent Univ, Elekt & Elekt Muhendisligi Bolumu, TR-06800 Ankara, Turkey
来源
2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU) | 2018年
关键词
Multi-armed bandit problems; risk minimization; ambulance redeployment; RELOCATION;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Ambulance redeployment comprises the problem of deploying ambulances to certain locations in order to minimize the arrival times to possible calls and plays a significant role in improving a country's emergency medical services and increasing the number of lives saved during an emergency. In this study, unlike the existing optimization methods in the literature, the problem is cast as a multi-armed bandit problem. Multi-armed bandit problems are a part of sequential online learning methods and utilized in maximizing a gain function (i.e. reward) when the reward distributions are unknown. In this study, in addition to the objective of maximizing rewards, the objective of minimizing the expected variance of rewards is also considered. The effect of risk taken by the system on average arrival times and number of calls responded on time is investigated. Ambulance redeployment is performed by a risk-averse multi-armed bandit algorithm on a data-driven simulator. As a result, it is shown that the algorithm which takes less risk (i.e. that minimizes the variance of response times) responds to more cases on time.
引用
收藏
页数:4
相关论文
共 50 条
  • [21] Ballooning multi-armed bandits
    Ghalme, Ganesh
    Dhamal, Swapnil
    Jain, Shweta
    Gujar, Sujit
    Narahari, Y.
    ARTIFICIAL INTELLIGENCE, 2021, 296
  • [22] Adaptive Data Depth via Multi-Armed Bandits
    Baharav, Tavor Z.
    Lai, Tze Leung
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [23] Residential HVAC Aggregation Based on Risk-averse Multi-armed Bandit Learning for Secondary Frequency Regulation
    Chen, Xinyi
    Hu, Qinran
    Shi, Qingxin
    Quan, Xiangjun
    Wu, Zaijun
    Li, Fangxing
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2020, 8 (06) : 1160 - 1167
  • [24] Residential HVAC Aggregation Based on Risk-averse Multi-armed Bandit Learning for Secondary Frequency Regulation
    Xinyi Chen
    Qinran Hu
    Qingxin Shi
    Xiangjun Quan
    Zaijun Wu
    Fangxing Li
    JournalofModernPowerSystemsandCleanEnergy, 2020, 8 (06) : 1160 - 1167
  • [25] Fast multiple sequence alignment via multi-armed bandits
    Mazooji, Kayvon
    Shomorony, Ilan
    BIOINFORMATICS, 2024, 40 : i328 - i336
  • [26] Designing Incentives for Crowdsourced Tasks via Multi-Armed Bandits
    Itoh, Akihiko
    Matsubara, Shigeo
    2016 IEEE INTERNATIONAL CONFERENCE ON AGENTS (IEEE ICA 2016), 2016, : 70 - 73
  • [27] ProtoBandit: Efficient Prototype Selection via Multi-Armed Bandits
    Chaudhuri, Arghya Roy
    Jawanpuria, Pratik
    Mishra, Bamdev
    Proceedings of Machine Learning Research, 2022, 189 : 169 - 184
  • [28] Transfer in Sequential Multi-armed Bandits via Reward Samples
    Rahul, N. R.
    Katewa, Vaibhav
    2024 EUROPEAN CONTROL CONFERENCE, ECC 2024, 2024, : 2083 - 2089
  • [29] Multi-armed bandits for performance marketing
    Gigli, Marco
    Stella, Fabio
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024,
  • [30] Lenient Regret for Multi-Armed Bandits
    Merlis, Nadav
    Mannor, Shie
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8950 - 8957