LSTTN: A Long-Short Term Transformer-based spatiotemporal neural network for traffic flow

被引:49
作者
Luo, Qinyao [1 ,2 ]
He, Silu [1 ]
Han, Xing [1 ]
Wang, Yuhan [1 ]
Li, Haifeng [1 ,2 ]
机构
[1] Cent South Univ, Sch Geosci & Infophys, 932 South Lushan Rd, Changsha 410083, Hunan, Peoples R China
[2] Xiangjiang Lab, 569 YueLu Ave, Changsha 410205, Hunan, Peoples R China
关键词
Traffic forecasting; Spatiotemporal modeling; Long-short term forecasting; Transformer; Mask Subseries Strategy; GRAPH CONVOLUTIONAL NETWORK; GCN;
D O I
10.1016/j.knosys.2024.111637
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Accurate traffic forecasting is a fundamental problem in intelligent transportation systems and learning longrange traffic representations with key information through spatiotemporal graph neural networks (STGNNs) is a basic assumption of current traffic flow prediction models. However, due to structural limitations, existing STGNNs can only utilize short-range traffic flow data; therefore, the models cannot adequately learn the complex trends and periodic features in traffic flow. Besides, it is challenging to extract the key temporal information from the long historical traffic series and obtain a compact representation. To solve the above problems, we propose a novel LSTTN (Long -Short Term Transformer -based Network) framework comprehensively considering the longand short-term features in historical traffic flow. First, we employ a masked subseries Transformer to infer the content of masked subseries from a small portion of unmasked subseries and their temporal context in a pretraining manner, forcing the model to efficiently learn compressed and contextual subseries temporal representations from long historical series. Then, based on the learned representations, long-term trend is extracted by using stacked 1D dilated convolution layers, and periodic features are extracted by dynamic graph convolution layers. For the difficulties in making time -step level prediction, LSTTN adopts a short-term trend extractor to learn fine-grained short-term temporal features. Finally, LSTTN fuses the long-term trend, periodic features and short-term features to obtain the prediction results. Experiments on four real -world datasets show that in 60 -minute -ahead long-term forecasting, the LSTTN model achieves a minimum improvement of 5.63% and a maximum improvement of 16.78% over baseline models. The source code is availble at https://github.com/GeoX-Lab/LSTTN.
引用
收藏
页数:12
相关论文
共 56 条
[1]  
Ba J, 2014, ACS SYM SER
[2]   A3T-GCN: Attention Temporal Graph Convolutional Network for Traffic Forecasting [J].
Bai, Jiandong ;
Zhu, Jiawei ;
Song, Yujiao ;
Zhao, Ling ;
Hou, Zhixiang ;
Du, Ronghua ;
Li, Haifeng .
ISPRS INTERNATIONAL JOURNAL OF GEO-INFORMATION, 2021, 10 (07)
[3]  
Bruna J, 2014, Arxiv, DOI [arXiv:1312.6203, DOI 10.48550/ARXIV.1312.6203]
[4]   Traffic transformer: Capturing the continuity and periodicity of time series for traffic forecasting [J].
Cai, Ling ;
Janowicz, Krzysztof ;
Mai, Gengchen ;
Yan, Bo ;
Zhu, Rui .
TRANSACTIONS IN GIS, 2020, 24 (03) :736-755
[5]   Predictions of Freeway Traffic Speeds and Volumes Using Vector Autoregressive Models [J].
Chandra, Srinivasa Ravi ;
Al-Deek, Haitham .
JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2009, 13 (02) :53-72
[6]   SELF-SUPERVISED LEARNING FOR FEW-SHOT IMAGE CLASSIFICATION [J].
Chen, Da ;
Chen, Yuefeng ;
Li, Yuhong ;
Mao, Feng ;
He, Yuan ;
Xue, Hui .
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, :1745-1749
[7]  
Defferrard M, 2016, ADV NEUR IN, V29
[8]  
Devlin J, 2019, Arxiv, DOI [arXiv:1810.04805, DOI 10.48550/ARXIV.1810.04805]
[9]  
Gilmer J, 2017, PR MACH LEARN RES, V70
[10]  
Guo SN, 2019, AAAI CONF ARTIF INTE, P922