A deep inverse reinforcement learning approach to route choice modeling with context-dependent rewards

被引:24
作者
Zhao, Zhan [1 ,2 ]
Liang, Yuebing [1 ]
机构
[1] Univ Hong Kong, Dept Urban Planning & Design, Hong Kong, Peoples R China
[2] Univ Hong Kong, Musketeers Fdn Inst Data Sci, Hong Kong, Peoples R China
关键词
Route choice modeling; Inverse reinforcement learning; Deep neural networks; Travel behavior; Trajectory data mining; RECURSIVE LOGIT MODEL;
D O I
10.1016/j.trc.2023.104079
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Route choice modeling is a fundamental task in transportation planning and demand forecasting. Classical methods generally adopt the discrete choice model (DCM) framework with linear utility functions and high-level route characteristics. While several recent studies have started to explore the applicability of deep learning for route choice modeling, they are all path-based with relatively simple model architectures and require choice set generation. Existing link-based models can capture the dynamic nature of link choices within the trip without the need for choice set generation, but still assume linear relationships and link-additive features. To address these issues, this study proposes a general deep inverse reinforcement learning (IRL) framework for link-based route choice modeling, which is capable of incorporating diverse features (of the state, action and trip context) and capturing complex relationships. Specifically, we adapt an adversarial IRL model to the route choice problem for efficient estimation of context-dependent reward functions without value iteration. Experiment results based on taxi GPS data from Shanghai, China validate the superior prediction performance of the proposed model over conventional DCMs and other imitation learning baselines, even for destinations unseen in the training data. Further analysis shows that the model exhibits competitive computational efficiency and reasonable interpretability. The proposed methodology provides a new direction for future development of route choice models. It is general and can be adaptable to other route choice problems across different modes and networks.
引用
收藏
页数:23
相关论文
共 45 条
  • [1] Abbeel P., 2004, P 21 INT C MACHINE L, P1
  • [2] Why did you predict that? Towards explainable artificial neural networks for travel demand analysis[J]. Alwosheel, Ahmad;van Cranenburgh, Sander;Chorus, Caspar G. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2021
  • [3] AN ALGORITHM FOR THE RANKING OF SHORTEST PATHS[J]. AZEVEDO, JA;COSTA, MEOS;MADEIRA, JJERS;MARTINS, EQV. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 1993(01)
  • [4] Ben-Akiva M., 1999, HDB TRANSPORTATION S, P5, DOI DOI 10.1007/978-1-4615-5203-1_2
  • [5] Multilayer feedforward networks for transportation mode choice analysis: An analysis and a comparison with random utility models[J]. Cantarella, GE;de Luca, S. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2005(02)
  • [6] TrajGAIL: Generating urban vehicle trajectories using generative adversarial imitation learning[J]. Choi, Seongjin;Kim, Jiwon;Yeo, Hwasoo. TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2021
  • [7] Doshi-Velez F, 2017, Arxiv, DOI arXiv:1702.08608
  • [8] Finn C, 2016, PR MACH LEARN RES, V48
  • [9] Finn C, 2016, Arxiv, DOI arXiv:1611.03852
  • [10] A link based network route choice model with unrestricted choice set[J]. Fosgerau, Mogens;Frejinger, Emma;Karlstrom, Anders. TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2013