Data-Centric Benchmarking of Neural Network Architectures for the Univariate Time Series Forecasting Task

被引:0
作者
Schlieper, Philipp [1 ]
Dombrowski, Mischa [1 ]
Nguyen, An [1 ]
Zanca, Dario [1 ]
Eskofier, Bjoern [1 ,2 ]
机构
[1] Friedrich Alexander Univ, Dept Artificial Intelligence Biomed Engn, D-91052 Erlangen, Germany
[2] Helmholtz Ctr Munich, German Res Ctr Environm Hlth, Inst AI Hlth, D-85764 Neuherberg, Germany
关键词
deep learning; time series; neural networks; model selection; data synthesis; univariate forecasting;
D O I
10.3390/forecast6030037
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Time series forecasting has witnessed a rapid proliferation of novel neural network approaches in recent times. However, performances in terms of benchmarking results are generally not consistent, and it is complicated to determine in which cases one approach fits better than another. Therefore, we propose adopting a data-centric perspective for benchmarking neural network architectures on time series forecasting by generating ad hoc synthetic datasets. In particular, we combine sinusoidal functions to synthesize univariate time series data for multi-input-multi-output prediction tasks. We compare the most popular architectures for time series, namely long short-term memory (LSTM) networks, convolutional neural networks (CNNs), and transformers, and directly connect their performance with different controlled data characteristics, such as the sequence length, noise and frequency, and delay length. Our findings suggest that transformers are the best architecture for dealing with different delay lengths. In contrast, for different noise and frequency levels and different sequence lengths, LSTM is the best-performing architecture by a significant amount. Based on our insights, we derive recommendations which allow machine learning (ML) practitioners to decide which architecture to apply, given the dataset's characteristics.
引用
收藏
页码:718 / 747
页数:30
相关论文
共 35 条
[1]  
Agarwal Kushagra, 2020, 2020 19th IEEE International Conference on Machine Learning and Applications (ICMLA), P859, DOI 10.1109/ICMLA51294.2020.00140
[2]  
Akiba T, 2019, Arxiv, DOI arXiv:1907.10902
[3]  
[Anonymous], 2006, Time Series Analysis and Its Applications With R Examples in R
[4]  
Bai SJ, 2018, Arxiv, DOI [arXiv:1803.01271, DOI 10.48550/ARXIV.1803.01271]
[5]   Benchmarking Attention-Based Interpretability of Deep Learning in Multivariate Time Series Predictions [J].
Baric, Domjan ;
Fumic, Petar ;
Horvatic, Davor ;
Lipic, Tomislav .
ENTROPY, 2021, 23 (02) :1-23
[6]   Libra: A Benchmark for Time Series Forecasting Methods [J].
Bauer, Andre ;
Zuefle, Marwin ;
Eismann, Simon ;
Grohmann, Johannes ;
Herbst, Nikolas ;
Kounev, Samuel .
PROCEEDINGS OF THE ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING (ICPE '21), 2021, :189-200
[7]  
Bergstra J., 2013, P 30 INT C MACHINE L, P115
[8]   FAST FOURIER TRANSFORM AND ITS APPLICATIONS [J].
COOLEY, JW ;
LEWIS, PAW ;
WELCH, PD .
IEEE TRANSACTIONS ON EDUCATION, 1969, E 12 (01) :27-&
[9]   DLIO: A Data-Centric Benchmark for Scientific Deep Learning Applications [J].
Devarajan, Hariharan ;
Zheng, Huihuo ;
Kougkas, Anthony ;
Sun, Xian-He ;
Vishwanath, Venkatram .
21ST IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING (CCGRID 2021), 2021, :81-91
[10]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1