Decision-making for Connected and Automated Vehicles in Chanllenging Traffic Conditions Using Imitation and Deep Reinforcement Learning

被引:3
作者
Hu, Jinchao [1 ]
Li, Xu [1 ]
Hu, Weiming [1 ]
Xu, Qimin [1 ]
Hu, Yue [1 ]
机构
[1] Southeast Univ, Sch Instrument Sci & Engn, Nanjing 210096, Peoples R China
关键词
Connected and automated vehicles (CAVs); Traffic safety; Decision-making; Imitation learning; Deep reinforcement learning;
D O I
10.1007/s12239-023-0128-0
中图分类号
TH [机械、仪表工业];
学科分类号
0802 ;
摘要
Decision-making is the "brain" of connected and automated vehicles (CAVs) and is vitally critical to the safety of CAVs. The most of driving data used to train the decision-making algorithms is collected in general traffic conditions. Existing decision-making methods are difficult to guarantee safety in challenging traffic conditions, namely severe congestion and accident ahead. In this context, a semi-supervised decision-making algorithm is proposed to improve the safety of CAVs in challenging traffic conditions. To be specific, we proposed the expert-generative adversarial imitation learning (E-GAIL) that integrates imitation learning and deep reinforcement learning. The proposed E-GAIL is deployed in roadside unit (RSU). In the first stage, the decision-making knowledge of the expert is imitated using the real-world data collected in general traffic conditions. In the second stage, the generator of E-GAIL is further reinforced and achieves self-learn decision-making in the simulator with challenging traffic conditions. The E-GAIL is tested in general and challenging traffic conditions. By comparing the evaluation metrics of time to collision (TTC), deceleration to avoid a crash (DRAC), space gap (SGAP) and time gap (TGAP), the E-GAIL greatly outperforms the state-of-the-art decision-making algorithms. Experimental results show that the E-GAIL not only make-decision for CAVs in general traffic conditions but also successfully enhances the safety of CAVs in challenging traffic conditions.
引用
收藏
页码:1589 / 1602
页数:14
相关论文
共 34 条
  • [1] Ahmedov H. B., 2021, arXiv
  • [2] A Survey on 3D Object Detection Methods for Autonomous Driving Applications
    Arnold, Eduardo
    Al-Jarrah, Omar Y.
    Dianati, Mehrdad
    Fallah, Saber
    Oxtoby, David
    Mouzakitis, Alex
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (10) : 3782 - 3795
  • [3] Florence P., 2022, P 5 C ROB LEARN CORL
  • [4] Proximal Policy Optimization With Policy Feedback
    Gu, Yang
    Cheng, Yuhu
    Chen, C. L. Philip
    Wang, Xuesong
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2022, 52 (07): : 4600 - 4610
  • [5] Safety performance measures: a comparison between microsimulation and observational data
    Guido, Giuseppe
    Astarita, Vittorio
    Giofre, Vincenzo
    Vitale, Alessandro
    [J]. STATE OF THE ART IN THE EUROPEAN QUANTITATIVE ORIENTED TRANSPORTATION AND LOGISTICS RESEARCH, 2011: 14TH EURO WORKING GROUP ON TRANSPORTATION & 26TH MINI EURO CONFERENCE & 1ST EUROPEAN SCIENTIFIC CONFERENCE ON AIR TRANSPORT, 2011, 20
  • [6] Fuel-Saving Control Strategy for Fuel Vehicles with Deep Reinforcement Learning and Computer Vision
    Han, Ling
    Liu, Guopeng
    Zhang, Hui
    Fang, Ruoyu
    Zhu, Changsheng
    [J]. INTERNATIONAL JOURNAL OF AUTOMOTIVE TECHNOLOGY, 2023, 24 (03) : 609 - 621
  • [7] Ho J, 2016, ADV NEUR IN, V29
  • [8] Combining Planning and Deep Reinforcement Learning in Tactical Decision Making for Autonomous Driving
    Hoel, Carl-Johan
    Driggs-Campbell, Katherine
    Wolff, Krister
    Laine, Leo
    Kochenderfer, Mykel J.
    [J]. IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2020, 5 (02): : 294 - 305
  • [9] Autonomous Vehicle Cut-In Algorithm for Lane-Merging Scenarios via Policy-Based Reinforcement Learning Nested Within Finite-State Machine
    Hwang, Seulbin
    Lee, Kibeom
    Jeon, Hyeongseok
    Kum, Dongsuk
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (10) : 17594 - 17606
  • [10] Deep Q-network-based multi-criteria decision-making framework for virtual simulation environment
    Jang, Hyeonjun
    Hao, Shujia
    Chu, Phuong Minh
    Sharma, Pradip Kumar
    Sung, Yunsick
    Cho, Kyungeun
    [J]. NEURAL COMPUTING & APPLICATIONS, 2021, 33 (17) : 10657 - 10671