Generative Adversarial Imitation Learning Based Bicycle Behaviors Simulation on Road Segments

被引:0
|
作者
Wei, Shuqiao [1 ]
Ni, Ying [1 ]
Sun, Jian [1 ]
Qiu, Hongtong [2 ]
机构
[1] Key Laboratory of Road and Traffic Engineering, the Ministry of Education, Tongji University, Shanghai,201804, China
[2] Traffic Management Research Institute of the Ministry of Public Security, Jiangsu, Wuxi,214151, China
来源
Jiaotong Yunshu Xitong Gongcheng Yu Xinxi/Journal of Transportation Systems Engineering and Information Technology | 2024年 / 24卷 / 04期
基金
中国国家自然科学基金;
关键词
Automobile driver simulators - Automobile testing - Homogenization method - Lagrange multipliers - Maneuverability - Reinforcement learning;
D O I
10.16097/j.cnki.1009-6744.2024.04.011
中图分类号
学科分类号
摘要
In order to accurately reproduce the interaction behavior of bicycles to meet the needs of autonomous driving simulation testing, a Position Reward Augmented Generative Adversarial Imitation Learning (PRA-GAIL) method is proposed. In urban roads, since the disturbance behavior is mainly generated by electric bicycles, electric bicycles are selected as the research object. In the constructed simulation environment, Generative Adversarial Imitation Learning (GAIL) is used to make the simulated trajectories approximate the real trajectories, while Position Reward and Lagrangian Constraint methods are added to solve the homogenization and uncontrollable behaviors of existing simulation methods. In the test set validation, the average displacement error of the GAIL and PRA-GAIL methods decreased by 61.7% and 65.8%, respectively, compared to the behavioral cloning method. In the behavioral performance validation, the KL divergence of acceleration distributions between simulation and reality was significantly reduced in PRA-GAIL compared to GAIL, and the percentage error of overtaking and illegal lane-changing behaviors decreased by 7.2% and 20.2%, respectively. Using the Lagrangian method to add constraints resulted in a 75.8% reduction in the number of agents with risky behavior compared to commonly used reward augmentation methods. In trajectory validation, in the simulation environment, the average displacement error of PRA-GAIL is reduced by 17.5% compared to GAIL. The resulting model realistically reproduces the overtaking maneuver space of cyclists. The results show that the method adopted in this paper is suitable for bicycle behavior simulation, the proposed modifications effectively enhance the simulation performance, and the obtained simulation model accurately reproduces the disturbance behavior of bicycles on road segments, which can be applied to automated vehicle simulation tests. © 2024 Science Press. All rights reserved.
引用
收藏
页码:105 / 115
相关论文
共 50 条
  • [1] Generative Adversarial Imitation Learning
    Ho, Jonathan
    Ermon, Stefano
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [2] A Survey of Imitation Learning Based on Generative Adversarial Nets
    Lin J.-H.
    Zhang Z.-Z.
    Jiang C.
    Hao J.-Y.
    Jisuanji Xuebao/Chinese Journal of Computers, 2020, 43 (02): : 326 - 351
  • [3] Ranking-Based Generative Adversarial Imitation Learning
    Shi, Zhipeng
    Zhang, Xuehe
    Fang, Yu
    Li, Changle
    Liu, Gangfeng
    Zhao, Jie
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (10): : 8967 - 8974
  • [4] Quantum generative adversarial imitation learning
    Xiao, Tailong
    Huang, Jingzheng
    Li, Hongjing
    Fan, Jianping
    Zeng, Guihua
    NEW JOURNAL OF PHYSICS, 2023, 25 (03):
  • [5] Combining Model-Based Controllers and Generative Adversarial Imitation Learning for Traffic Simulation
    Chen, Haonan
    Ji, Tianchen
    Liu, Shuijing
    Driggs-Campbell, Katherine
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 1698 - 1704
  • [6] Deterministic generative adversarial imitation learning
    Zuo, Guoyu
    Chen, Kexin
    Lu, Jiahao
    Huang, Xiangsheng
    NEUROCOMPUTING, 2020, 388 : 60 - 69
  • [7] GACS: Generative Adversarial Imitation Learning Based on Control Sharing
    Huaiwei SI
    Guozhen TAN
    Dongyu LI
    Yanfei PENG
    Journal of Systems Science and Information, 2023, 11 (01) : 78 - 93
  • [8] Imitation Learning for Playing Shogi Based on Generative Adversarial Networks
    Wan, Shanchuan
    Kaneko, Tomoyuki
    2017 CONFERENCE ON TECHNOLOGIES AND APPLICATIONS OF ARTIFICIAL INTELLIGENCE (TAAI), 2017, : 92 - 95
  • [9] A Bayesian Approach to Generative Adversarial Imitation Learning
    Jeon, Wonseok
    Seo, Seokin
    Kim, Kee-Eung
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] Modeling Car-Following Behaviors and Driving Styles with Generative Adversarial Imitation Learning
    Zhou, Yang
    Fu, Rui
    Wang, Chang
    Zhang, Ruibin
    SENSORS, 2020, 20 (18) : 1 - 20