Human-like autonomous car-following model with deep reinforcement learning

被引:336
作者
Zhu, Meixin [1 ,2 ,3 ]
Wang, Xuesong [1 ,2 ]
Wang, Yinhai [3 ]
机构
[1] Minist Educ, Key Lab Rd & Traff Engn, Shanghai 201804, Peoples R China
[2] Tongji Univ, Sch Transportat Engn, Shanghai 201804, Peoples R China
[3] Univ Washington, Dept Civil & Environm Engn, Seattle, WA 98195 USA
基金
美国国家科学基金会;
关键词
Autonomous car following; Human-like driving planning; Deep reinforcement learning; Naturalistic driving study; Deep deterministic policy gradient; NETWORKS;
D O I
10.1016/j.trc.2018.10.024
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
This study proposes a framework for human-like autonomous car-following planning based on deep reinforcement learning (deep RL). Historical driving data are fed into a simulation environment where an RL agent learns from trial and error interactions based on a reward function that signals how much the agent deviates from the empirical data. Through these interactions, an optimal policy, or car-following model that maps in a human-like way from speed, relative speed between a lead and following vehicle, and inter-vehicle spacing to acceleration of a following vehicle is finally obtained. The model can be continuously updated when more data are fed in. Two thousand car-following periods extracted from the 2015 Shanghai Naturalistic Driving Study were used to train the model and compare its performance with that of traditional and recent data driven car-following models. As shown by this study's results, a deep deterministic policy gradient car-following model that uses disparity between simulated and observed speed as the reward function and considers a reaction delay of 1 s, denoted as DDPGvRT, can reproduce human-like car following behavior with higher accuracy than traditional and recent data-driven car-following models. Specifically, the DDPGvRT model has a spacing validation error of 18% and speed validation error of 5%, which are less than those of other models, including the intelligent driver model, models based on locally weighted regression, and conventional neural network-based models. Moreover, the DDPGvRT demonstrates good capability of generalization to various driving situations and can adapt to different drivers by continuously learning. This study demonstrates that reinforcement learning methodology can offer insight into driver behavior and can contribute to the development of human-like autonomous driving algorithms and traffic-flow models.
引用
收藏
页码:348 / 368
页数:21
相关论文
共 50 条
[21]   Human-like Object Grasping and Relocation for an Anthropomorphic Robotic Hand with Natural Hand Pose Priors in Deep Reinforcement Learning [J].
Valarezo Anazco, Edwin ;
Lopez, Patricio Rivera ;
Park, Hyemin ;
Park, Nahyeon ;
Oh, Jiheon ;
Lee, Sangmin ;
Byun, Kyungmin ;
Kim, Tae-Seong .
ICRSA 2019: 2019 2ND INTERNATIONAL CONFERENCE ON ROBOT SYSTEMS AND APPLICATIONS, 2019, :46-50
[22]   Control Schemes for Autonomous Car-following Systems with two Classical Compensators [J].
Zhu, Wen-Xing ;
Zhang, Li-Dong .
ASIAN JOURNAL OF CONTROL, 2020, 22 (01) :168-181
[23]   A Deep Reinforcement Learning Based Approach for Autonomous Overtaking [J].
Li, Xiaoxiang ;
Qiu, Xinyou ;
Wang, Jian ;
Shen, Yuan .
2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2020,
[24]   Deep Reinforcement Learning for Autonomous Internet of Things: Model, Applications and Challenges [J].
Lei, Lei ;
Tan, Yue ;
Zheng, Kan ;
Liu, Shiwen ;
Zhang, Kuan ;
Shen, Xuemin .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (03) :1722-1760
[25]   Physics-informed deep reinforcement learning-based integrated two-dimensional car-following control strategy for connected automated vehicles [J].
Shi, Haotian ;
Zhou, Yang ;
Wu, Keshu ;
Chen, Sikai ;
Ran, Bin ;
Nie, Qinghui .
KNOWLEDGE-BASED SYSTEMS, 2023, 269
[26]   Deep Robot Sketching: An application of Deep Q-Learning Networks for human-like sketching [J].
Fernandez-Fernandez, Raul ;
Victores, Juan G. ;
Balaguer, Carlos .
COGNITIVE SYSTEMS RESEARCH, 2023, 81 :57-63
[27]   A reinforcement learning model for autonomous vehicles with realistic car simulation in urban using Unity [J].
Mamdouh, Abdulla ;
Farouk, Mohamed ;
Khater, Hatem ;
Hassan, Yasser .
ENGINEERING RESEARCH EXPRESS, 2024, 6 (04)
[28]   HDRLM3D: A Deep Reinforcement Learning-Based Model with Human-like Perceptron and Policy for Crowd Evacuation in 3D Environments [J].
Zhang, Dong ;
Li, Wenhang ;
Gong, Jianhua ;
Huang, Lin ;
Zhang, Guoyong ;
Shen, Shen ;
Liu, Jiantao ;
Ma, Haonan .
ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2022, 11 (04)
[29]   Autonomous Vehicle Driving Path Control with Deep Reinforcement Learning [J].
Tiong, Teckchai ;
Saad, Ismail ;
Teo, Kenneth Tze Kin ;
bin Lago, Herwansyah .
2023 IEEE 13TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE, CCWC, 2023, :84-92
[30]   Deep Reinforcement Learning for Autonomous Driving: A Survey [J].
Kiran, B. Ravi ;
Sobh, Ibrahim ;
Talpaert, Victor ;
Mannion, Patrick ;
Al Sallab, Ahmad A. ;
Yogamani, Senthil ;
Perez, Patrick .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (06) :4909-4926