Modeling framework of human driving behavior based on Deep Maximum Entropy Inverse Reinforcement Learning

被引:0
|
作者
Wang, Yongjie [1 ]
Niu, Yuchen [1 ]
Xiao, Mei [1 ]
Zhu, Wenying [1 ]
You, Xinshang [2 ]
机构
[1] Changan Univ, Sch Transportat Engn, Xian 710064, Peoples R China
[2] Hebei Univ Sci & Technol, Sch Econ & Management, Shijiazhuang 050018, Peoples R China
基金
中国国家自然科学基金;
关键词
Human driving behavior; Autonomous vehicle; Occluded pedestrian; Inverse reinforcement learning; Reinforcement learning; PERFORMANCE;
D O I
10.1016/j.physa.2024.130052
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Driving behavior modeling is extremely crucial for designing safe, intelligent, and personalized autonomous driving systems. In this paper, a modeling framework based on Markov Decision Processes (MDPs) is introduced that emulates drivers' decision-making processes. The framework combines the Deep Maximum Entropy Inverse Reinforcement Learning (Deep MEIRL) and a reinforcement learning algorithm-proximal strategy optimization (PPO). A neural network structure is customized for Deep MEIRL, which uses the velocity of the ego vehicle, the pedestrian position, the velocity of surrounding vehicles, the lateral distance, the surrounding vehicles' type, and the distance to the crosswalk to recover the nonlinear reward function. The dataset of drone-based video footage is collected in Xi'an (China) to train and validate the framework. The outcomes demonstrate that Deep MEIRL-PPO outperforms traditional modeling frameworks (Maximum Entropy Inverse Reinforcement Learning (MEIRL)- PPO) in modeling and predicting human driving behavior. Specifically, in predicting human driving behavior, Deep MEIRL-PPO outperforms MEIRL-PPO by 50.71% and 43.90% on the basis of the MAE and HD, respectively. Furthermore, it is discovered that Deep MEIRL-PPO accurately learns the behavior of human drivers avoiding potential conflicts when lines of sight are occluded. This research can contribute to aiding self-driving vehicles in learning human driving behavior and avoiding unforeseen risks.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Car-Following Behavior Modeling With Maximum Entropy Deep Inverse Reinforcement Learning
    Nan, Jiangfeng
    Deng, Weiwen
    Zhang, Ruzheng
    Zhao, Rui
    Wang, Ying
    Ding, Juan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (02): : 3998 - 4010
  • [2] Driving Behavior Modeling Using Naturalistic Human Driving Data With Inverse Reinforcement Learning
    Huang, Zhiyu
    Wu, Jingda
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 10239 - 10251
  • [3] Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning With Application to Autonomous Driving
    Wu, Zheng
    Sun, Liting
    Zhan, Wei
    Yang, Chenyu
    Tomizuka, Masayoshi
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (04) : 5355 - 5362
  • [4] AdaBoost maximum entropy deep inverse reinforcement learning with truncated gradient
    Song, Li
    Li, Dazi
    Wang, Xiao
    Xu, Xin
    INFORMATION SCIENCES, 2022, 602 : 328 - 350
  • [5] Simulation of Vehicle Interaction Behavior in Merging Scenarios: A Deep Maximum Entropy-Inverse Reinforcement Learning Method Combined With Game Theory
    Li, Wenli
    Qiu, Fanke
    Li, Lingxi
    Zhang, Yinan
    Wang, Kan
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 1079 - 1093
  • [6] Revisiting Maximum Entropy Inverse Reinforcement Learning: New Perspectives and Algorithm
    Snoswell, Aaron J.
    Singh, Surya P. N.
    Ye, Nan
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 241 - 249
  • [7] Adaptive generative adversarial maximum entropy inverse reinforcement learning
    Song, Li
    Li, Dazi
    Xu, Xin
    INFORMATION SCIENCES, 2025, 695
  • [8] A Maximum Entropy Deep Reinforcement Learning Neural Tracker
    Balaram, Shafa
    Arulkumaran, Kai
    Dai, Tianhong
    Bharath, Anil Anthony
    MACHINE LEARNING IN MEDICAL IMAGING (MLMI 2019), 2019, 11861 : 400 - 408
  • [9] Conditional Predictive Behavior Planning With Inverse Reinforcement Learning for Human-Like Autonomous Driving
    Huang, Zhiyu
    Liu, Haochen
    Wu, Jingda
    Lv, Chen
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2023, 24 (07) : 7244 - 7258
  • [10] Predicting Pedestrian Crossing Behavior at Unsignalized Mid-Block Crosswalks Using Maximum Entropy Deep Inverse Reinforcement Learning
    Wang, Yongjie
    Niu, Yuchen
    Zhu, Wenying
    Chen, Wenqiang
    Li, Qiong
    Wang, Tao
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (05) : 3685 - 3698