Performance Analysis of Long Short-Term Memory Predictive Neural Networks on Time Series Data

被引:14
作者
Bolboaca, Roland [1 ]
Haller, Piroska [1 ]
机构
[1] George Emil Palade Univ Med Pharm Sci & Technol Ta, Fac Engn & Informat Technol, Targu Mures 540139, Romania
关键词
long short-term memory (LSTM); recurrent neural network (RNN); teacher forcing; prediction; performance analysis; benchmarking; machine learning; Tennessee Eastman process; time series; LSTM; MODEL;
D O I
10.3390/math11061432
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Long short-term memory neural networks have been proposed as a means of creating accurate models from large time series data originating from various fields. These models can further be utilized for prediction, control, or anomaly-detection algorithms. However, finding the optimal hyperparameters to maximize different performance criteria remains a challenge for both novice and experienced users. Hyperparameter optimization algorithms can often be a resource-intensive and time-consuming task, particularly when the impact of the hyperparameters on the performance of the neural network is not comprehended or known. Teacher forcing denotes a procedure that involves feeding the ground truth output from the previous time-step as input to the current time-step during training, while during testing feeding back the predicted values. This paper presents a comprehensive examination of the impact of hyperparameters on long short-term neural networks, with and without teacher forcing, on prediction performance. The study includes testing long short-term memory neural networks, with two variations of teacher forcing, in two prediction modes, using two configurations (i.e., multi-input single-output and multi-input multi-output) on a well-known chemical process simulation dataset. Furthermore, this paper demonstrates the applicability of a long short-term memory neural network with a modified teacher forcing approach in a process state monitoring system. Over 100,000 experiments were conducted with varying hyperparameters and in multiple neural network operation modes, revealing the direct impact of each tested hyperparameter on the training and testing procedures.
引用
收藏
页数:35
相关论文
共 80 条
[61]   Time series forecasting of petroleum production using deep LSTM recurrent networks [J].
Sagheer, Alaa ;
Kotb, Mostafa .
NEUROCOMPUTING, 2019, 323 (203-213) :203-213
[62]   DeepAR: Probabilistic forecasting with autoregressive recurrent networks [J].
Salinas, David ;
Flunkert, Valentin ;
Gasthaus, Jan ;
Januschowski, Tim .
INTERNATIONAL JOURNAL OF FORECASTING, 2020, 36 (03) :1181-1191
[63]  
Schmidt F, 2019, Arxiv, DOI [arXiv:1910.00292, 10.48550/arXiv.1910.00292]
[64]  
Schrauwen B., 2007, P 15 EUROPEAN S ARTI, P471, DOI [10.1007/978-3-540-74690-448, DOI 10.1007/978-3-540-74690-448]
[65]  
Shailaja K., 2018, 2018 Second International Conference on Electronics, Communication and Aerospace Technology (ICECA), P910, DOI 10.1109/ICECA.2018.8474918
[67]  
Siami-Namini S, 2019, IEEE INT CONF BIG DA, P3285, DOI 10.1109/BigData47090.2019.9005997
[68]   A Comparison of ARIMA and LSTM in Forecasting Time Series [J].
Siami-Namini, Sima ;
Tavakoli, Neda ;
Namin, Akbar Siami .
2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, :1394-1401
[69]  
Staudemeyer R.C., 2019, ARXIVT
[70]  
Taigman Y, 2018, Arxiv, DOI arXiv:1707.06588