Bidirectional Spatial-Temporal Adaptive Transformer for Urban Traffic Flow Forecasting

被引:74
作者
Chen, Changlu [1 ]
Liu, Yanbin [2 ]
Chen, Ling [1 ]
Zhang, Chengqi [1 ]
机构
[1] Univ Technol Sydney, Australian Artificial Intelligence Inst, Sydney, NSW 2007, Australia
[2] Univ Western Australia, UWA Ctr Med Res, Perth, WA 6009, Australia
关键词
Dynamic halting mechanism; spatial-temporal; transformer; urban traffic forecasting; FUTURE; MEMORY;
D O I
10.1109/TNNLS.2022.3183903
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Urban traffic forecasting is the cornerstone of the intelligent transportation system (ITS). Existing methods focus on spatial-temporal dependency modeling, while two intrinsic properties of the traffic forecasting problem are overlooked. First, the complexity of diverse forecasting tasks is nonuniformly distributed across various spaces (e.g., suburb versus downtown) and times (e.g., rush hour versus off-peak). Second, the recollection of past traffic conditions is beneficial to the prediction of future traffic conditions. Based on these properties, we propose a bidirectional spatial-temporal adaptive transformer (Bi-STAT) for accurate traffic forecasting. Bi-STAT adopts an encoder-decoder architecture, where both the encoder and the decoder maintain a spatial-adaptive transformer and a temporal-adaptive transformer structure. Inspired by the first property, each transformer is designed to dynamically process the traffic streams according to their task complexities. Specifically, we realize this by the recurrent mechanism with a novel dynamic halting module (DHM). Each transformer performs iterative computation with shared parameters until DHM emits a stopping signal. Motivated by the second property, Bi-STAT utilizes one decoder to perform the present -> past recollection task and the other decoder to perform the present -> future prediction task. The recollection task supplies complementary information to assist and regularize the prediction task for a better generalization. Through extensive experiments, we show the effectiveness of each module in Bi-STAT and demonstrate the superiority of Bi-STAT over the state-of-the-art baselines on four benchmark datasets.
引用
收藏
页码:6913 / 6925
页数:13
相关论文
共 50 条
[1]  
Ba J. L., 2016, CoRR
[2]  
Bai L, 2020, ADV NEUR IN, V33
[3]  
Bai L, 2019, PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P1981
[4]   Traffic transformer: Capturing the continuity and periodicity of time series for traffic forecasting [J].
Cai, Ling ;
Janowicz, Krzysztof ;
Mai, Gengchen ;
Yan, Bo ;
Zhu, Rui .
TRANSACTIONS IN GIS, 2020, 24 (03) :736-755
[5]   Emerging Properties in Self-Supervised Vision Transformers [J].
Caron, Mathilde ;
Touvron, Hugo ;
Misra, Ishan ;
Jegou, Herve ;
Mairal, Julien ;
Bojanowski, Piotr ;
Joulin, Armand .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9630-9640
[6]  
Chen CHW, 2001, AIP CONF PROC, V584, P96, DOI 10.1063/1.1405589
[7]  
Chen K., 2021, PROC INT C MACH LEAR, P1684
[8]   Learning on Attribute-Missing Graphs [J].
Chen, Xu ;
Chen, Siheng ;
Yao, Jiangchao ;
Zheng, Huangjie ;
Zhang, Ya ;
Tsang, Ivor W. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (02) :740-757
[9]  
Chen YZ, 2021, PR MACH LEARN RES, V139
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171