Take an Irregular Route: Enhance the Decoder of Time-Series Forecasting Transformer

被引:4
作者
Shen, Li [1 ]
Wei, Yuning [1 ]
Wang, Yangzhu [2 ]
Qiu, Huaxin [2 ]
机构
[1] Beihang Univ, Sch Automat Sci & Elect Engn, Beijing 100191, Peoples R China
[2] Beijing Univ, Flying Coll, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; neural network; time-series forecasting; transformer;
D O I
10.1109/JIOT.2023.3341099
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the development of Internet of Things (IoT) systems, precise long-term forecasting method is requisite for decision makers to evaluate current statuses and formulate future policies. Currently, transformer and MLP are two paradigms for deep time-series forecasting and the former one is more prevailing in virtue of its exquisite attention mechanism and encoder-decoder architecture. However, data scientists seem to be more willing to dive into the research of encoder, leaving decoder unconcerned. Some researchers even adopt linear projections in lieu of the decoder to reduce the complexity. We argue that both extracting the features of input sequence and seeking the relations of input and prediction sequence, which are respective functions of encoder and decoder, are of paramount significance. Motivated from the success of FPN in CV field, we propose FPPformer to utilize bottom-up and top-down architectures, respectively, in encoder and decoder to build the full and rational hierarchy. The cutting-edge patchwise attention is exploited and further developed with the combination, whose format is also different in encoder and decoder, of revamped elementwise attention in this work. Extensive experiments with six state-of-the-art baselines on twelve benchmarks verify the promising performances of FPPformer and the importance of elaborately devising decoder in time-series forecasting transformer. The source code is released in https://github.com/OrigamiSL/FPPformer.
引用
收藏
页码:14344 / 14356
页数:13
相关论文
共 48 条
[1]   A Bias and Variance Analysis for Multistep-Ahead Time Series Forecasting [J].
Ben Taieb, Souhaib ;
Atiya, Amir F. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (01) :62-76
[2]   Deep Learning for Time Series Forecasting: Tutorial and Literature Survey [J].
Benidis, Konstantinos ;
Rangapuram, Syama Sundar ;
Flunkert, Valentin ;
Wang, Yuyang ;
Maddix, Danielle ;
Turkmen, Caner ;
Gasthaus, Jan ;
Bohlke-Schneider, Michael ;
Salinas, David ;
Stella, Lorenzo ;
Aubet, Francois-Xavier ;
Callot, Laurent ;
Januschowski, Tim .
ACM COMPUTING SURVEYS, 2023, 55 (06)
[3]  
bgc-jena, 2023, Max-planck-institut fuer biogeochemie-wetterdaten (n.d.)
[4]  
BOX GEP, 1974, ROY STAT SOC C-APP, V23, P158, DOI 10.2307/2346997
[5]  
Challu C, 2023, AAAI CONF ARTIF INTE, P6989
[6]  
Chen SC, 2023, PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, P3532
[7]  
Cirstea RG, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P1994
[8]   TSMixer: Lightweight MLP-Mixer Model for Multivariate Time Series Forecasting [J].
Ekambaram, Vijay ;
Jati, Arindam ;
Nguyen, Nam ;
Sinthong, Phanwadee ;
Kalagnanam, Jayant .
PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, :459-469
[9]  
Fan W, 2023, AAAI CONF ARTIF INTE, P7522
[10]   TIME-SERIES ANALYSIS - FORECASTING AND CONTROL - BOX,GEP AND JENKINS,GM [J].
GEURTS, M .
JOURNAL OF MARKETING RESEARCH, 1977, 14 (02) :269-269