Predicting Citywide Passenger Demand via Reinforcement Learning from Spatio-Temporal Dynamics

被引:4
|
作者
Ning, Xiaodong [1 ]
Yao, Lina [1 ]
Wang, Xianzhi [2 ]
Benatallah, Boualem [1 ]
Salim, Flora [3 ]
Haghighi, Pari Delir [4 ]
机构
[1] Univ New South Wales, Sydney, NSW, Australia
[2] Univ Technol Sydney, Sydney, NSW, Australia
[3] RMIT Univ, Melbourne, Vic, Australia
[4] Monash Univ, Clayton, Vic, Australia
来源
PROCEEDINGS OF THE 15TH EAI INTERNATIONAL CONFERENCE ON MOBILE AND UBIQUITOUS SYSTEMS: COMPUTING, NETWORKING AND SERVICES (MOBIQUITOUS 2018) | 2018年
关键词
Reinforcement Learning; spatial-temporal dynamics; passenger demand prediction;
D O I
10.1145/3286978.3286991
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The global urbanization imposes unprecedented pressure on urban infrastructure and public resources. The population explosion has made it challenging to satisfy the daily needs of urban residents. 'Smart City' is a solution that utilizes different types of data collection sensors to help manage assets and resources intelligently and more efficiently. Under the Smart City umbrella, the primary research initiative in improving the efficiency of car-hailing services is to predict the citywide passenger demand to address the imbalance between the demand and supply. However, predicting the passenger demand requires analysis on various data such as historical passenger demand, crowd outflow, and weather information, and it remains challenging to discover the latent relationships among these data. To address this challenge, we propose to improve the passenger demand prediction via learning the salient spatialtemporal dynamics within a reinforcement learning framework. Our model employs an information selection mechanism to focus on the most distinctive data in historical observations. This mechanism can automatically adjust the information zone according to the prediction performance to find the optimal choice. It also ensures the prediction model to take full advantage of the available data by introducing the positive and excluding the negative correlations. We have conducted experiments on a large-scale real-world dataset that covers 1.5 million people in a major city in China. The results show our model outperforms state-of-the-art and a series of baselines by a large margin.
引用
收藏
页码:19 / 28
页数:10
相关论文
共 50 条
  • [21] Imitation Learning of Neural Spatio-Temporal Point Processes
    Zhu, Shixiang
    Li, Shuang
    Peng, Zhigang
    Xie, Yao
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (11) : 5391 - 5402
  • [22] MOTION LEARNING USING SPATIO-TEMPORAL NEURAL NETWORK
    Yusoff, Nooraini
    Kabir-Ahmad, Farzana
    Jemili, Mohamad-Farif
    JOURNAL OF INFORMATION AND COMMUNICATION TECHNOLOGY-MALAYSIA, 2020, 19 (02): : 207 - 223
  • [23] Multiobjective optimization deep reinforcement learning for dependent task scheduling based on spatio-temporal fusion graph neural network
    Wang, Zhi
    Zhan, Wenhan
    Duan, Hancong
    Huang, Hualong
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 148
  • [24] Spatio-temporal task pricing for shared electric micro-mobility battery-swapping platform with reinforcement learning
    Kim, Minjeong
    Moon, Ilkyeong
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2025, 63 (04) : 1473 - 1494
  • [25] DANet: A spatio-temporal dynamics and Detail Aware Network for video prediction
    Huang, Huilin
    Guan, YePeng
    NEUROCOMPUTING, 2024, 598
  • [26] TEMPORAL LINK PREDICTION VIA REINFORCEMENT LEARNING
    Tao, Ye
    Li, Ying
    Wu, Zhonghai
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3470 - 3474
  • [27] Contextual spatio-temporal graph representation learning for reinforced human mobility mining
    Gao, Qiang
    Zhou, Fan
    Zhong, Ting
    Trajcevski, Goce
    Yang, Xin
    Li, Tianrui
    INFORMATION SCIENCES, 2022, 606 : 230 - 249
  • [28] From predicting to decision making: Reinforcement learning in biomedicine
    Liu, Xuhan
    Zhang, Jun
    Hou, Zhonghuai
    Yang, Yi Isaac
    Gao, Yi Qin
    WILEY INTERDISCIPLINARY REVIEWS-COMPUTATIONAL MOLECULAR SCIENCE, 2024, 14 (04)
  • [29] DynaSTI: Dynamics modeling with sequential temporal information for reinforcement learning in Atari
    Kim, Jaehoon
    Lee, Young Jae
    Kwak, Mingu
    Park, Young Joon
    Kim, Seoung Bum
    KNOWLEDGE-BASED SYSTEMS, 2024, 299
  • [30] Reinforcement learning: computing the temporal difference of values via distinct corticostriatal pathways
    Morita, Kenji
    Morishima, Mieko
    Sakai, Katsuyuki
    Kawaguchi, Yasuo
    TRENDS IN NEUROSCIENCES, 2012, 35 (08) : 457 - 467