A deep inverse reinforcement learning approach to route choice modeling with context-dependent rewards

被引:22
|
作者
Zhao, Zhan [1 ,2 ]
Liang, Yuebing [1 ]
机构
[1] Univ Hong Kong, Dept Urban Planning & Design, Hong Kong, Peoples R China
[2] Univ Hong Kong, Musketeers Fdn Inst Data Sci, Hong Kong, Peoples R China
关键词
Route choice modeling; Inverse reinforcement learning; Deep neural networks; Travel behavior; Trajectory data mining; RECURSIVE LOGIT MODEL;
D O I
10.1016/j.trc.2023.104079
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Route choice modeling is a fundamental task in transportation planning and demand forecasting. Classical methods generally adopt the discrete choice model (DCM) framework with linear utility functions and high-level route characteristics. While several recent studies have started to explore the applicability of deep learning for route choice modeling, they are all path-based with relatively simple model architectures and require choice set generation. Existing link-based models can capture the dynamic nature of link choices within the trip without the need for choice set generation, but still assume linear relationships and link-additive features. To address these issues, this study proposes a general deep inverse reinforcement learning (IRL) framework for link-based route choice modeling, which is capable of incorporating diverse features (of the state, action and trip context) and capturing complex relationships. Specifically, we adapt an adversarial IRL model to the route choice problem for efficient estimation of context-dependent reward functions without value iteration. Experiment results based on taxi GPS data from Shanghai, China validate the superior prediction performance of the proposed model over conventional DCMs and other imitation learning baselines, even for destinations unseen in the training data. Further analysis shows that the model exhibits competitive computational efficiency and reasonable interpretability. The proposed methodology provides a new direction for future development of route choice models. It is general and can be adaptable to other route choice problems across different modes and networks.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Learning context-dependent choice functions
    Pfannschmidt, Karlson
    Gupta, Pritha
    Haddenhorst, Bjoern
    Huellermeier, Eyke
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2022, 140 : 116 - 155
  • [2] A Route Choice Model with Context-Dependent Value of Time
    Xu, Hongli
    Yang, Hai
    Zhou, Jing
    Yin, Yafeng
    TRANSPORTATION SCIENCE, 2017, 51 (02) : 536 - 548
  • [3] Individual Versus Difference Rewards on Reinforcement Learning for Route Choice
    Grunitzki, Ricardo
    Ramos, Gabriel de O.
    Bazzan, Ana L. C.
    2014 BRAZILIAN CONFERENCE ON INTELLIGENT SYSTEMS (BRACIS), 2014, : 253 - 258
  • [4] Enhancing Joint Behavior Modeling with Route Choice using Adversarial Inverse Reinforcement Learning
    Ogawa, Daichi
    Hato, Eiji
    2023 IEEE 26TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS, ITSC, 2023, : 4750 - 4755
  • [5] Context-dependent outcome encoding in human reinforcement learning
    Palminteri, Stefano
    Lebreton, Mael
    CURRENT OPINION IN BEHAVIORAL SCIENCES, 2021, 41 : 144 - 151
  • [6] Context-Dependent Deep Learning for Affective Computing
    Suresh, Varsha
    2022 10TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS, ACIIW, 2022,
  • [7] Context-dependent extinction learning emerging from raw sensory inputs: a reinforcement learning approach
    Thomas Walther
    Nicolas Diekmann
    Sandhiya Vijayabaskaran
    José R. Donoso
    Denise Manahan-Vaughan
    Laurenz Wiskott
    Sen Cheng
    Scientific Reports, 11
  • [8] Context-dependent extinction learning emerging from raw sensory inputs: a reinforcement learning approach
    Walther, Thomas
    Diekmann, Nicolas
    Vijayabaskaran, Sandhiya
    Donoso, Jose R.
    Manahan-Vaughan, Denise
    Wiskott, Laurenz
    Cheng, Sen
    SCIENTIFIC REPORTS, 2021, 11 (01)
  • [9] Skills Transfer Across Dissimilar Robots by Learning Context-Dependent Rewards
    Malekzadeh, Milad S.
    Bruno, Danilo
    Calinon, Sylvain
    Nanayakkara, Thrishantha
    Caldwell, Darwin G.
    2013 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2013, : 1746 - 1751
  • [10] Testing models of context-dependent outcome encoding in reinforcement learning
    Hayes, William M.
    Wedell, Douglas H.
    COGNITION, 2023, 230