Toward Learning Human-Like, Safe and Comfortable Car-Following Policies With a Novel Deep Reinforcement Learning Approach

被引:8
作者
Yavas, M. Ugur [1 ]
Kumbasar, Tufan [2 ]
Ure, Nazim Kemal [3 ]
机构
[1] Istanbul Tech Univ, Dept Mechatron Engn, TR-34469 Istanbul, Turkiye
[2] Istanbul Tech Univ, Dept Control & Automat Engn, TR-34469 Istanbul, Turkiye
[3] Istanbul Tech Univ, Artificial Intelligence & Data Sci Res Ctr, TR-34469 Istanbul, Turkiye
关键词
Vehicles; Cruise control; Behavioral sciences; Advanced driver assistance systems; Tuning; Safety; Heuristic algorithms; Reinforcement learning; Deep learning; Adaptive cruise control; reinforcement learning; deep learning; naturalistic driving; advanced driving assistance systems; ADAPTIVE CRUISE CONTROL; MODEL;
D O I
10.1109/ACCESS.2023.3245831
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we present an advanced adaptive cruise control (ACC) concept powered by Deep Reinforcement Learning (DRL) that generates safe, human-like, and comfortable car-following policies. Unlike the current trend in developing DRL-based ACC systems, we propose defining the action space of the DRL agent with discrete actions rather than continuous ones, since human drivers never set the throttle/brake pedal level to be actuated, but rather the required change of the current pedal levels. Through this human-like throttle-brake manipulation representation, we also define explicit actions for holding (keeping the last action) and coasting (no action), which are usually omitted as actions in ACC systems. Moreover, based on the investigation of a real-world driving dataset, we cast a novel reward function that is easy to interpret and personalized. The proposed reward enforces the agent to learn stable and safe actions, while also encouraging the holding and coasting actions, just like a human driver would. The proposed discrete action DRL agent is trained with action masking, and the reward terms are completely derived from the real-world dataset collected from a human driver. We present exhaustive comparative results to show the advantages of the proposed DRL approach in both simulation and scenarios extracted from real-world driving. We clearly show that the proposed policy imitates human driving significantly better and handles complex driving situations, such as cut-ins and cut-outs, implicitly, in comparison with a DRL agent trained with a widely-used reward function proposed for ACC, a model predictive control structure, and traditional car-following approaches.
引用
收藏
页码:16843 / 16854
页数:12
相关论文
共 29 条
[1]   Driver Modeling Through Deep Reinforcement Learning and Behavioral Game Theory [J].
Albaba, Berat Mert ;
Yildiz, Yildiray .
IEEE TRANSACTIONS ON CONTROL SYSTEMS TECHNOLOGY, 2022, 30 (02) :885-892
[2]  
[Anonymous], 2009, NGSIM NEXT GEN SIM
[3]  
Ayres TJ, 2001, 2001 IEEE INTELLIGENT TRANSPORTATION SYSTEMS - PROCEEDINGS, P826, DOI 10.1109/ITSC.2001.948767
[4]   Personalized Speed Planning Algorithm Using a Statistical Driver Model in Car-following Situations [J].
Baek, Seung Eon ;
Kim, Hak Su ;
Han, Manbae .
INTERNATIONAL JOURNAL OF AUTOMOTIVE TECHNOLOGY, 2022, 23 (03) :829-840
[5]  
Busoniu L, 2010, AUTOM CONTROL ENG SE, P1, DOI 10.1201/9781439821091-f
[6]   Personalized Adaptive Cruise Control Based on Online Driving Style Recognition Technology and Model Predictive Control [J].
Gao, Bingzhao ;
Cai, Kunyang ;
Qu, Ting ;
Hu, Yunfeng ;
Chen, Hong .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (11) :12482-12496
[7]  
Hart F., 2021, arXiv
[8]  
Horgan D, 2018, Arxiv, DOI arXiv:1803.00933
[9]   A car-following model considering asymmetric driving behavior based on long short-term memory neural networks [J].
Huang, Xiuling ;
Sun, Jie ;
Sun, Jian .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2018, 95 :346-362
[10]  
Kanervisto A, 2020, IEEE CONF COMPU INTE, P479, DOI [10.1109/cog47356.2020.9231687, 10.1109/CoG47356.2020.9231687]