Temporal Chain Network With Intuitive Attention Mechanism for Long-Term Series Forecasting

被引:8
作者
Zhang, Zhen [1 ,2 ]
Han, Yongming [1 ,2 ]
Ma, Bo [3 ]
Liu, Min [4 ,5 ]
Geng, Zhiqiang [1 ,2 ]
机构
[1] Beijing Univ Chem Technol, Coll Informat Sci & Technol, Beijing 100029, Peoples R China
[2] Minist Educ China, Engn Res Ctr Intelligent PSE, Beijing 100029, Peoples R China
[3] Beijing Univ Chem Technol, Coll Mech & Elect Engn, Beijing 100029, Peoples R China
[4] Beijing Univ Chem Technol, Beijing Adv Innovat Ctr Soft Matter Sci & Engn, Beijing 100029, Peoples R China
[5] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Attention mechanism; neural networks; permutation-invariance; time series forecasting; transformer; TIME-SERIES; NEURAL-NETWORK; TRANSFORMER;
D O I
10.1109/TIM.2023.3322508
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Long-term series forecasting (LTSF) plays an important role in real-world applications in the economy, the weather, and the industrial process. At present, many transformer-based methods have made promising progress. However, the nature of the permutation-invariant self-attention mechanism inevitably results in temporal information loss, which hinders the prediction performance of the transformer-based LTSF methods. Therefore, this article proposed a novel temporal chain network (TCNet) with an intuitive attention mechanism for LTSF. Based on the chain forward propagation structure of time series, a one-way chain graph neural network (GNN) is constructed to avoid the permutation-invariance of the self-attention mechanism. Meanwhile, based on the natural forgetting mechanism of time series, the prior intuitive attention is proposed as the edge weight (attention) for information propagation and then the series model is obtained by the GNN. Furthermore, the proposed method achieves the LTSF task by linear forecasting of the trend component and nonlinear forecasting of the seasonal component of the series. Extensive experiments on five benchmarks and a real-world chemical process dataset are conducted to demonstrate the effectiveness of the TCNet. Comparison experiment results show that the TCNet achieves state-of-the-art results compared to current baselines and reduces average prediction error by 9.52% and 5.97% on transformer-based and nontransformer-based multivariate LTSF baselines, respectively. Moreover, the temporal information loss for the TCNet due to the permutation-invariance of the self-attention mechanism is not the main reason hindering transformer-based LTSF methods, whose bottleneck comes mainly from the complex architecture of the decoder.
引用
收藏
页数:13
相关论文
共 38 条
[1]   Stock Price Prediction Using the ARIMA Model [J].
Adebiyi, Ayodele A. ;
Adewumi, Aderemi O. ;
Ayo, Charles K. .
2014 UKSIM-AMSS 16TH INTERNATIONAL CONFERENCE ON COMPUTER MODELLING AND SIMULATION (UKSIM), 2014, :106-112
[2]  
Ba JL, 2016, arXiv
[3]  
Bai SJ, 2018, Arxiv, DOI [arXiv:1803.01271, DOI 10.48550/ARXIV.1803.01271]
[4]   LSTM-MSNet: Leveraging Forecasts on Sets of Related Time Series With Multiple Seasonal Patterns [J].
Bandara, Kasun ;
Bergmeir, Christoph ;
Hewamalage, Hansika .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (04) :1586-1599
[5]   Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach [J].
Bandara, Kasun ;
Bergmeir, Christoph ;
Smyl, Slawek .
EXPERT SYSTEMS WITH APPLICATIONS, 2020, 140
[6]  
Bertasius G, 2021, PR MACH LEARN RES, V139
[7]   Accurate medium-range global weather forecasting with 3D neural networks [J].
Bi, Kaifeng ;
Xie, Lingxi ;
Zhang, Hengheng ;
Chen, Xin ;
Gu, Xiaotao ;
Tian, Qi .
NATURE, 2023, 619 (7970) :533-+
[8]  
Brody S., 2022, P 10 INT C LEARN REP
[9]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[10]   A hybrid neural network and ARIMA model for water quality time series prediction [J].
Faruk, Durdu Oemer .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2010, 23 (04) :586-594