Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
[41]   Learning Automated Driving in Complex Intersection Scenarios Based on Camera Sensors: A Deep Reinforcement Learning Approach [J].
Li, Guofa ;
Lin, Siyan ;
Li, Shen ;
Qu, Xingda .
IEEE SENSORS JOURNAL, 2022, 22 (05) :4687-4696
[42]   Individualized Driving Intention Prediction With Inverse Reinforcement Learning [J].
Liu, Siqi ;
Li, Xinyang ;
Chen, Jiansheng ;
Guo, Chenghao ;
Wu, Jiehui ;
Luo, Qifeng ;
Ma, Huimin .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (06) :8125-8139
[43]   Deep reinforcement learning framework and algorithms integrated with cognitive behavior models [J].
Chen H. ;
Li J.-X. ;
Huang J. ;
Wang C. ;
Liu Q. ;
Zhang Z.-J. .
Kongzhi yu Juece/Control and Decision, 2023, 38 (11) :3209-3218
[44]   A synergistic reinforcement learning-based framework design in driving automation [J].
Qi, Yuqiong ;
Hu, Yang ;
Wu, Haibin ;
Li, Shen ;
Ye, Xiaochun ;
Fan, Dongrui .
COMPUTERS & ELECTRICAL ENGINEERING, 2022, 101
[45]   Individual Human Behavior Identification Using an Inverse Reinforcement Learning Method [J].
Inga, Jairo ;
Koepf, Florian ;
Flad, Michael ;
Hohmann, Soeren .
2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, :99-104
[46]   Deep inverse reinforcement learning for structural evolution of small molecules [J].
Agyemang, Brighter ;
Wu, Wei-Ping ;
Addo, Daniel ;
Kpiebaareh, Michael Y. ;
Nanor, Ebenezer ;
Haruna, Charles Roland .
BRIEFINGS IN BIOINFORMATICS, 2021, 22 (04)
[47]   Score-based Inverse Reinforcement Learning [J].
El Asri, Layla ;
Piot, Bilal ;
Geist, Matthieu ;
Laroche, Romain ;
Pietquin, Olivier .
AAMAS'16: PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2016, :457-465
[48]   Path tracking control based on Deep reinforcement learning in Autonomous driving [J].
Jiang, Le ;
Wang, Yafei ;
Wang, Lin ;
Wu, Jingkai .
2019 3RD CONFERENCE ON VEHICLE CONTROL AND INTELLIGENCE (CVCI), 2019, :414-419
[49]   Deep Reinforcement Learning for Autonomous Driving Based on Safety Experience Replay [J].
Huang, Xiaohan ;
Cheng, Yuhu ;
Yu, Qiang ;
Wang, Xuesong .
IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (06) :2070-2084
[50]   Recent advances in reinforcement learning-based autonomous driving behavior planning: A survey [J].
Wu, Jingda ;
Huang, Chao ;
Huang, Hailong ;
Lv, Chen ;
Wang, Yuntong ;
Wang, Fei-Yue .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2024, 164