Predicting Citywide Passenger Demand via Reinforcement Learning from Spatio-Temporal Dynamics

被引:4
|
作者
Ning, Xiaodong [1 ]
Yao, Lina [1 ]
Wang, Xianzhi [2 ]
Benatallah, Boualem [1 ]
Salim, Flora [3 ]
Haghighi, Pari Delir [4 ]
机构
[1] Univ New South Wales, Sydney, NSW, Australia
[2] Univ Technol Sydney, Sydney, NSW, Australia
[3] RMIT Univ, Melbourne, Vic, Australia
[4] Monash Univ, Clayton, Vic, Australia
来源
PROCEEDINGS OF THE 15TH EAI INTERNATIONAL CONFERENCE ON MOBILE AND UBIQUITOUS SYSTEMS: COMPUTING, NETWORKING AND SERVICES (MOBIQUITOUS 2018) | 2018年
关键词
Reinforcement Learning; spatial-temporal dynamics; passenger demand prediction;
D O I
10.1145/3286978.3286991
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The global urbanization imposes unprecedented pressure on urban infrastructure and public resources. The population explosion has made it challenging to satisfy the daily needs of urban residents. 'Smart City' is a solution that utilizes different types of data collection sensors to help manage assets and resources intelligently and more efficiently. Under the Smart City umbrella, the primary research initiative in improving the efficiency of car-hailing services is to predict the citywide passenger demand to address the imbalance between the demand and supply. However, predicting the passenger demand requires analysis on various data such as historical passenger demand, crowd outflow, and weather information, and it remains challenging to discover the latent relationships among these data. To address this challenge, we propose to improve the passenger demand prediction via learning the salient spatialtemporal dynamics within a reinforcement learning framework. Our model employs an information selection mechanism to focus on the most distinctive data in historical observations. This mechanism can automatically adjust the information zone according to the prediction performance to find the optimal choice. It also ensures the prediction model to take full advantage of the available data by introducing the positive and excluding the negative correlations. We have conducted experiments on a large-scale real-world dataset that covers 1.5 million people in a major city in China. The results show our model outperforms state-of-the-art and a series of baselines by a large margin.
引用
收藏
页码:19 / 28
页数:10
相关论文
共 50 条
  • [31] Stable reinforcement learning via temporal competition between LTP and LTD traces
    Marco A Huertas
    Sarah Schwettmann
    Alfredo Kirkwood
    Harel Shouval
    BMC Neuroscience, 15 (Suppl 1)
  • [32] Reinforcement learning based on intrinsic motivation and temporal abstraction via transformation invariance
    Masuyama, Gakuto
    Yamashita, Atsushi
    Asama, Hajime
    Masuyama, G. (masuyama@robot.t.u-tokyo.ac.jp), 1600, Japan Society of Mechanical Engineers (79): : 289 - 303
  • [33] Learning to make auto-scaling decisions with heterogeneous spot and on-demand instances via reinforcement learning
    Lin, Liduo
    Pan, Li
    Liu, Shijun
    INFORMATION SCIENCES, 2022, 614 (480-496) : 480 - 496
  • [34] Uncovering the spatio-temporal dynamics of value-based decision-making in the human brain: a combined fMRI - EEG study
    Larsen, Tobias
    O'Doherty, John P.
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY B-BIOLOGICAL SCIENCES, 2014, 369 (1655)
  • [35] PATROL: A Velocity Control Framework for Autonomous Vehicle via Spatial-Temporal Reinforcement Learning
    Xu, Zhi
    Liu, Shuncheng
    Wu, Ziniu
    Chen, Xu
    Zeng, Kai
    Zheng, Kai
    Su, Han
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2271 - 2280
  • [36] Closing the Dynamics Gap via Adversarial and Reinforcement Learning for High-Speed Racing
    Niu, Jingyu
    Hu, Yu
    Li, Wei
    Huang, Guangyan
    Han, Yinhe
    Li, Xiaowei
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [37] A Modular Simulation Platform for Training Robots via Deep Reinforcement Learning and Multibody Dynamics
    Benatti, Simone
    Tasora, Alessandro
    Fusai, Dario
    Mangoni, Dario
    PROCEEDINGS OF 2019 3RD INTERNATIONAL CONFERENCE ON AUTOMATION, CONTROL AND ROBOTS (ICACR 2019), 2018, : 7 - 11
  • [38] Robust Temporal Link Prediction in Dynamic Complex Networks via Stable Gated Models With Reinforcement Learning
    Yang, Liping
    Liu, Hongbo
    Sun, Daoqiang
    McLoone, Sean
    Liu, Kai
    Chen, C. L. Philip
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 12
  • [39] Exploring electron beam induced atomic assembly via reinforcement learning in a molecular dynamics environment *
    Vasudevan, Rama K.
    Ghosh, Ayana
    Ziatdinov, Maxim
    Kalinin, Sergei, V
    NANOTECHNOLOGY, 2022, 33 (11)
  • [40] A reinforcement learning-based demand response strategy designed from the Aggregator's perspective
    Oh, Seongmun
    Jung, Jaesung
    Onen, Ahmet
    Lee, Chul-Ho
    FRONTIERS IN ENERGY RESEARCH, 2022, 10