Human-like autonomous car-following model with deep reinforcement learning

被引:333
|
作者
Zhu, Meixin [1 ,2 ,3 ]
Wang, Xuesong [1 ,2 ]
Wang, Yinhai [3 ]
机构
[1] Minist Educ, Key Lab Rd & Traff Engn, Shanghai 201804, Peoples R China
[2] Tongji Univ, Sch Transportat Engn, Shanghai 201804, Peoples R China
[3] Univ Washington, Dept Civil & Environm Engn, Seattle, WA 98195 USA
基金
美国国家科学基金会;
关键词
Autonomous car following; Human-like driving planning; Deep reinforcement learning; Naturalistic driving study; Deep deterministic policy gradient; NETWORKS;
D O I
10.1016/j.trc.2018.10.024
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
This study proposes a framework for human-like autonomous car-following planning based on deep reinforcement learning (deep RL). Historical driving data are fed into a simulation environment where an RL agent learns from trial and error interactions based on a reward function that signals how much the agent deviates from the empirical data. Through these interactions, an optimal policy, or car-following model that maps in a human-like way from speed, relative speed between a lead and following vehicle, and inter-vehicle spacing to acceleration of a following vehicle is finally obtained. The model can be continuously updated when more data are fed in. Two thousand car-following periods extracted from the 2015 Shanghai Naturalistic Driving Study were used to train the model and compare its performance with that of traditional and recent data driven car-following models. As shown by this study's results, a deep deterministic policy gradient car-following model that uses disparity between simulated and observed speed as the reward function and considers a reaction delay of 1 s, denoted as DDPGvRT, can reproduce human-like car following behavior with higher accuracy than traditional and recent data-driven car-following models. Specifically, the DDPGvRT model has a spacing validation error of 18% and speed validation error of 5%, which are less than those of other models, including the intelligent driver model, models based on locally weighted regression, and conventional neural network-based models. Moreover, the DDPGvRT demonstrates good capability of generalization to various driving situations and can adapt to different drivers by continuously learning. This study demonstrates that reinforcement learning methodology can offer insight into driver behavior and can contribute to the development of human-like autonomous driving algorithms and traffic-flow models.
引用
收藏
页码:348 / 368
页数:21
相关论文
共 50 条
  • [1] Toward Learning Human-Like, Safe and Comfortable Car-Following Policies With a Novel Deep Reinforcement Learning Approach
    Yavas, M. Ugur
    Kumbasar, Tufan
    Ure, Nazim Kemal
    IEEE ACCESS, 2023, 11 : 16843 - 16854
  • [2] Human-like speed modeling for autonomous vehicles during car-following at intersections
    Tawfeek, Mostafa H.
    CANADIAN JOURNAL OF CIVIL ENGINEERING, 2022, 49 (02) : 255 - 264
  • [3] Driver Car-Following Model Based on Deep Reinforcement Learning
    Guo J.
    Li W.
    Luo Y.
    Chen T.
    Li K.
    Guo, Jinghua (guojh@xmu.edu.cn), 1600, SAE-China (43): : 571 - 579
  • [4] Dynamic Car-following Model Calibration with Deep Reinforcement Learning
    Naing, Htet
    Cai, Wentong
    Wu, Tiantian
    Yu, Liang
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 959 - 966
  • [5] Estimation of Varying Reaction Times with RNN and Application to Human-like Autonomous Car-following Modeling
    Ma, Lijing
    Qu, Shiru
    Zhang, Junxi
    Zhang, Xiangzhou
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (09) : 725 - 730
  • [6] Human-like car-following model for autonomous vehicles considering the cut-in behavior of other vehicles in mixed traffic
    Fu, Rui
    Li, Zhen
    Sun, Qinyu
    Wang, Chang
    ACCIDENT ANALYSIS AND PREVENTION, 2019, 132
  • [7] Deep Reinforcement Learning Car-Following Model Considering Longitudinal and Lateral Control
    Qin, Pinpin
    Tan, Hongyun
    Li, Hao
    Wen, Xuguang
    SUSTAINABILITY, 2022, 14 (24)
  • [8] A Combined Reinforcement Learning and Model Predictive Control for Car-Following Maneuver of Autonomous Vehicles
    Liwen Wang
    Shuo Yang
    Kang Yuan
    Yanjun Huang
    Hong Chen
    Chinese Journal of Mechanical Engineering, 36
  • [9] A Combined Reinforcement Learning and Model Predictive Control for Car-Following Maneuver of Autonomous Vehicles
    Wang, Liwen
    Yang, Shuo
    Yuan, Kang
    Huang, Yanjun
    Chen, Hong
    CHINESE JOURNAL OF MECHANICAL ENGINEERING, 2023, 36 (01)
  • [10] A Combined Reinforcement Learning and Model Predictive Control for Car-Following Maneuver of Autonomous Vehicles
    Liwen Wang
    Shuo Yang
    Kang Yuan
    Yanjun Huang
    Hong Chen
    Chinese Journal of Mechanical Engineering, 2023, (03) : 330 - 340