Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Learning Behavior Styles with Inverse Reinforcement Learning
    Lee, Seong Jae
    popovic, Zoran
    [J]. ACM TRANSACTIONS ON GRAPHICS, 2010, 29 (04):
  • [32] A Bayesian Approach forQuantifying Data Scarcity when Modeling Human Behavior via Inverse Reinforcement Learning
    Hossain, Tahera
    Shen, Wanggang
    Antar, Anindya
    Prabhudesai, Snehal
    Inoue, Sozo
    Huan, Xun
    Banovic, Nikola
    [J]. ACM TRANSACTIONS ON COMPUTER-HUMAN INTERACTION, 2023, 30 (01)
  • [33] Policy-Based Reinforcement Learning for Assortative Matching in Human Behavior Modeling
    Deng, Ou
    Jin, Qun
    [J]. DIGITAL HUMAN MODELING AND APPLICATIONS IN HEALTH, SAFETY, ERGONOMICS AND RISK MANAGEMENT, DHM 2023, PT II, 2023, 14029 : 378 - 391
  • [34] Inverse Reinforcement Learning Based on Behaviors of a Learning Agent
    Sakurai, Shunsuke
    Oba, Shigeyuki
    Ishii, Shin
    [J]. NEURAL INFORMATION PROCESSING, PT I, 2015, 9489 : 724 - 732
  • [35] Learning Automated Driving in Complex Intersection Scenarios Based on Camera Sensors: A Deep Reinforcement Learning Approach
    Li, Guofa
    Lin, Siyan
    Li, Shen
    Qu, Xingda
    [J]. IEEE SENSORS JOURNAL, 2022, 22 (05) : 4687 - 4696
  • [36] Deep reinforcement learning framework and algorithms integrated with cognitive behavior models
    Chen H.
    Li J.-X.
    Huang J.
    Wang C.
    Liu Q.
    Zhang Z.-J.
    [J]. Kongzhi yu Juece/Control and Decision, 2023, 38 (11): : 3209 - 3218
  • [37] A synergistic reinforcement learning-based framework design in driving automation
    Qi, Yuqiong
    Hu, Yang
    Wu, Haibin
    Li, Shen
    Ye, Xiaochun
    Fan, Dongrui
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2022, 101
  • [38] Individual Human Behavior Identification Using an Inverse Reinforcement Learning Method
    Inga, Jairo
    Koepf, Florian
    Flad, Michael
    Hohmann, Soeren
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 99 - 104
  • [39] Enhancing Pedestrian Route Choice Models Through Maximum-Entropy Deep Inverse Reinforcement Learning With Individual Covariates (MEDIRL-IC)
    Li, Boyang
    Zhang, Wenjia
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, : 20446 - 20463
  • [40] Deep inverse reinforcement learning for structural evolution of small molecules
    Agyemang, Brighter
    Wu, Wei-Ping
    Addo, Daniel
    Kpiebaareh, Michael Y.
    Nanor, Ebenezer
    Haruna, Charles Roland
    [J]. BRIEFINGS IN BIOINFORMATICS, 2021, 22 (04)