A deep inverse reinforcement learning approach to route choice modeling with context-dependent rewards

被引:22
|
作者
Zhao, Zhan [1 ,2 ]
Liang, Yuebing [1 ]
机构
[1] Univ Hong Kong, Dept Urban Planning & Design, Hong Kong, Peoples R China
[2] Univ Hong Kong, Musketeers Fdn Inst Data Sci, Hong Kong, Peoples R China
关键词
Route choice modeling; Inverse reinforcement learning; Deep neural networks; Travel behavior; Trajectory data mining; RECURSIVE LOGIT MODEL;
D O I
10.1016/j.trc.2023.104079
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Route choice modeling is a fundamental task in transportation planning and demand forecasting. Classical methods generally adopt the discrete choice model (DCM) framework with linear utility functions and high-level route characteristics. While several recent studies have started to explore the applicability of deep learning for route choice modeling, they are all path-based with relatively simple model architectures and require choice set generation. Existing link-based models can capture the dynamic nature of link choices within the trip without the need for choice set generation, but still assume linear relationships and link-additive features. To address these issues, this study proposes a general deep inverse reinforcement learning (IRL) framework for link-based route choice modeling, which is capable of incorporating diverse features (of the state, action and trip context) and capturing complex relationships. Specifically, we adapt an adversarial IRL model to the route choice problem for efficient estimation of context-dependent reward functions without value iteration. Experiment results based on taxi GPS data from Shanghai, China validate the superior prediction performance of the proposed model over conventional DCMs and other imitation learning baselines, even for destinations unseen in the training data. Further analysis shows that the model exhibits competitive computational efficiency and reasonable interpretability. The proposed methodology provides a new direction for future development of route choice models. It is general and can be adaptable to other route choice problems across different modes and networks.
引用
收藏
页数:23
相关论文
共 50 条
  • [21] Inverse Reinforcement Learning with Simultaneous Estimation of Rewards and Dynamics
    Herman, Michael
    Gindele, Tobias
    Wagner, Joerg
    Schmitt, Felix
    Burgard, Wolfram
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 102 - 110
  • [22] Learning How Pedestrians Navigate: A Deep Inverse Reinforcement Learning Approach
    Fahad, Muhammad
    Chen, Zhuo
    Guo, Yi
    2018 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2018, : 819 - 826
  • [23] Is Context-Based Choice due to Context-Dependent Preferences?
    Kobi Kriesler
    Shmuel Nitzan
    Theory and Decision, 2008, 64 : 65 - 80
  • [24] System for context-dependent user modeling
    Nurmi, Petteri
    Salden, Alfons
    Lau, Sian Lun
    Suomela, Jukka
    Sutterer, Michael
    Millerat, Jean
    Martin, Miquel
    Lagerspetz, Eemil
    Poortinga, Remco
    ON THE MOVE TO MEANINGFUL INTERNET SYSTEMS 2006: OTM 2006 WORKSHOPS, PT 2, PROCEEDINGS, 2006, 4278 : 1894 - 1903
  • [25] Is context-based choice due to context-dependent preferences?
    Kriesler, Kobi
    Nitzan, Shmuel
    THEORY AND DECISION, 2008, 64 (01) : 65 - 80
  • [26] Eigentriphones for Context-Dependent Acoustic Modeling
    Ko, Tom
    Mak, Brian
    IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2013, 21 (06): : 1285 - 1294
  • [27] Limitations of Recursive Logit for Inverse Reinforcement Learning of Bicycle Route Choice Behavior in Amsterdam
    Koch, Thomas
    Dugundji, Elenna
    12TH INTERNATIONAL CONFERENCE ON AMBIENT SYSTEMS, NETWORKS AND TECHNOLOGIES (ANT) / THE 4TH INTERNATIONAL CONFERENCE ON EMERGING DATA AND INDUSTRY 4.0 (EDI40) / AFFILIATED WORKSHOPS, 2021, 184 : 492 - 499
  • [28] Multi-Objective Deep Inverse Reinforcement Learning through Direct Weights and Rewards Estimation
    Kishikawa, Daiko
    Arai, Sachiyo
    2022 61ST ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS (SICE), 2022, : 122 - 127
  • [29] Context-dependent olfactory learning in an insect
    Matsumoto, Y
    Mizunami, M
    LEARNING & MEMORY, 2004, 11 (03) : 288 - 293
  • [30] Context-dependent learning and causal structure
    Gershman, Samuel J.
    PSYCHONOMIC BULLETIN & REVIEW, 2017, 24 (02) : 557 - 565