Cooperative Highway Work Zone Merge Control Based on Reinforcement Learning in a Connected and Automated Environment

被引:28
作者
Ren, Tianzhu [1 ]
Xie, Yuanchang [2 ]
Jiang, Liming [2 ]
机构
[1] Amazon, Seattle, WA USA
[2] Univ Massachusetts, Dept Civil & Environm Engn, Lowell, MA 01854 USA
基金
美国国家科学基金会;
关键词
Vehicles;
D O I
10.1177/0361198120935873
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Given the aging infrastructure and the anticipated growing number of highway work zones in the U.S.A., it is important to investigate work zone merge control, which is critical for improving work zone safety and capacity. This paper proposes and evaluates a novel highway work zone merge control strategy based on cooperative driving behavior enabled by artificial intelligence. The proposed method assumes that all vehicles are fully automated, connected, and cooperative. It inserts two metering zones in the open lane to make space for merging vehicles in the closed lane. In addition, each vehicle in the closed lane learns how to adjust its longitudinal position optimally to find a safe gap in the open lane using an off-policy soft actor critic reinforcement learning (RL) algorithm, considering its surrounding traffic conditions. The learning results are captured in convolutional neural networks and used to control individual vehicles in the testing phase. By adding the metering zones and taking the locations, speeds, and accelerations of surrounding vehicles into account, cooperation among vehicles is implicitly considered. This RL-based model is trained and evaluated using a microscopic traffic simulator. The results show that this cooperative RL-based merge control significantly outperforms popular strategies such as late merge and early merge in terms of both mobility and safety measures. It also performs better than a strategy assuming all vehicles are equipped with cooperative adaptive cruise control.
引用
收藏
页码:363 / 374
页数:12
相关论文
共 33 条
  • [1] [Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
  • [2] Beacher A. G., 2004, Research Report No VTRC05-R6
  • [3] Bojarski Mariusz, 2016, arXiv
  • [4] Bouton M, 2019, IEEE INT C INTELL TR, P3441, DOI [10.1109/ITSC.2019.8916924, 10.1109/itsc.2019.8916924]
  • [5] A Survey of Monte Carlo Tree Search Methods
    Browne, Cameron B.
    Powley, Edward
    Whitehouse, Daniel
    Lucas, Simon M.
    Cowling, Peter I.
    Rohlfshagen, Philipp
    Tavener, Stephen
    Perez, Diego
    Samothrakis, Spyridon
    Colton, Simon
    [J]. IEEE TRANSACTIONS ON COMPUTATIONAL INTELLIGENCE AND AI IN GAMES, 2012, 4 (01) : 1 - 43
  • [6] Chen X., 2017, 96 ANN M TRANSP RES
  • [7] Cooperative Adaptive Cruise Control: A Reinforcement Learning Approach
    Desjardins, Charles
    Chaib-draa, Brahim
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2011, 12 (04) : 1248 - 1260
  • [8] Dudek C. L., 1984, NOTES WORK ZONE CAPA
  • [9] Endsley MR, 2017, J COGN ENG DECIS MAK, V11, P225, DOI 10.1177/1555343417695197
  • [10] Fares A, 2014, IEEE INT CONF CON AU, P1226, DOI 10.1109/ICCA.2014.6871097