Model-Free Learning of Corridor Clearance: A Near-Term Deployment Perspective

被引:2
|
作者
Suo, Dajiang [1 ]
Jayawardana, Vindula [2 ,3 ]
Wu, Cathy [4 ,5 ]
机构
[1] Arizona State Univ, Polytech Sch, Mesa, AZ 85212 USA
[2] Massachusetts Inst Technol MIT, Lab Informat & Decis Syst, Cambridge, MA 02139 USA
[3] Massachusetts Inst Technol MIT, Dept Elect Engn & Comp Sci, Cambridge, MA 02139 USA
[4] Massachusetts Inst Technol MIT, Dept Civil & Environm Engn, Lab Informat & Decis Syst, Cambridge, MA 02139 USA
[5] Massachusetts Inst Technol MIT, Inst Data Syst & Soc, Cambridge, MA 02139 USA
关键词
Connected and automated vehicles; emergency vehicle corridor clearance; mixed autonomy; intelligent transportation systems; shock wave theory; deep reinforcement learning; EMERGENCY VEHICLES; TIME; SYSTEM;
D O I
10.1109/TITS.2023.3344473
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
An emerging public health application of connected and automated vehicle (CAV) technologies is to reduce response times of emergency medical service (EMS) by indirectly coordinating traffic. Therefore, in this work we study the CAV-assisted corridor clearance for EMS vehicles from a short term deployment perspective. Existing research on this topic often overlooks the impact of EMS vehicle disruptions on regular traffic, assumes 100% CAV penetration, relies on real-time traffic signal timing data and queue lengths at intersections, and makes various assumptions about traffic settings when deriving optimal model-based CAV control strategies. However, these assumptions pose significant challenges for near-term deployment and limit the real-world applicability of such methods. To overcome these challenges and enhance real-world applicability in near-term, we propose a model-free approach employing deep reinforcement learning (DRL) for designing CAV control strategies, showing its reduced overhead in designing and greater scalability and performance compared to model-based methods. Our qualitative analysis highlights the complexities of designing scalable EMS corridor clearance controllers for diverse traffic settings in which DRL controller provides ease of design compared to the model-based methods. In numerical evaluations, the model-free DRL controller outperforms the model-based counterpart by improving traffic flow and even improving EMS travel times in scenarios when a single CAV is present. Across 19 considered settings, the learned DRL controller excels by 25% in reducing the travel time in six instances, achieving an average improvement of 9%. These findings underscore the potential and promise of model-free DRL strategies in advancing EMS response and traffic flow coordination, with a focus on practical near-term deployment.
引用
收藏
页码:4833 / 4848
页数:16
相关论文
共 50 条
  • [1] Near-term deployment of carbon capture and sequestration from biorefineries in the United States
    Sanchez, Daniel L.
    Johnson, Nils
    McCoy, Sean T.
    Turner, Peter A.
    Mach, Katharine J.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2018, 115 (19) : 4875 - 4880
  • [2] Model-Free Reinforcement Learning Algorithms: A Survey
    Calisir, Sinan
    Pehlivanoglu, Meltem Kurt
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [3] A novel Markov model for near-term railway delay prediction
    Xu, Jin
    Wang, Weiqi
    Gao, Zheming
    Luo, Haochen
    Wu, Qian
    COMPUTERS & INDUSTRIAL ENGINEERING, 2023, 181
  • [4] An Adaptive Model-Free Control Method for Metro Train Based on Deep Reinforcement Learning
    Lai, Wenzhu
    Chen, Dewang
    Huang, Yunhu
    Huang, Benzun
    ADVANCES IN NATURAL COMPUTATION, FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY, ICNC-FSKD 2022, 2023, 153 : 263 - 273
  • [5] Model-based and model-free learning strategies for wet clutch control
    Dutta, Abhishek
    Zhong, Yu
    Depraetere, Bruno
    Van Vaerenbergh, Kevin
    Ionescu, Clara
    Wyns, Bart
    Pinte, Gregory
    Nowe, Ann
    Swevers, Jan
    De Keyser, Robin
    MECHATRONICS, 2014, 24 (08) : 1008 - 1020
  • [6] Measurement-Free Fault-Tolerant Quantum Error Correction in Near-Term Devices
    Heussen, Sascha
    Locher, David F.
    Mueller, Markus
    INTERNATIONAL JOURNAL OF ONLINE PEDAGOGY AND COURSE DESIGN, 2024, 5 (01)
  • [7] Can model-free reinforcement learning explain deontological moral judgments?
    Ayars, Alisabeth
    COGNITION, 2016, 150 : 232 - 242
  • [8] Fusion of Microgrid Control With Model-Free Reinforcement Learning: Review and Vision
    She, Buxin
    Li, Fangxing
    Cui, Hantao
    Zhang, Jingqiu
    Bo, Rui
    IEEE TRANSACTIONS ON SMART GRID, 2023, 14 (04) : 3232 - 3245
  • [9] Seasonal difference in temporal transferability of an ecological model: near-term predictions of lemming outbreak abundances
    Kleiven, Eivind Flittie
    Henden, John-Andre
    Ims, Rolf Anker
    Yoccoz, Nigel Gilles
    SCIENTIFIC REPORTS, 2018, 8
  • [10] Learning explainable task-relevant state representation for model-free deep reinforcement learning
    Zhao, Tingting
    Li, Guixi
    Zhao, Tuo
    Chen, Yarui
    Xie, Ning
    Niu, Gang
    Sugiyama, Masashi
    NEURAL NETWORKS, 2024, 180