A PLS-based pruning algorithm for simplified long-short term memory neural network in time series prediction

被引:22
作者
Li, Wenjing
Wang, Xiaoxiao
Han, Honggui
Qiao, Junfei
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
[2] Beijing Artificial Intelligence Inst, Beijing 100124, Peoples R China
[3] Minist Educ, Engn Res Ctr Intelligence Percept & Autonomous Co, Beijing 100124, Peoples R China
[4] Beijing Key Lab Computat Intelligence & Intellige, Beijing 100124, Peoples R China
[5] Beijing Lab Intelligent Environm Protect, Beijing 100124, Peoples R China
关键词
Time series prediction; Internal structure simplification; Partial least squares (PLS) regression; Pruning algorithm; Hidden layer size; LSTM; MODEL;
D O I
10.1016/j.knosys.2022.109608
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As an extensively used model for time series prediction, the Long-Short Term Memory (LSTM) neural network suffers from shortcomings such as high computational cost and large memory requirement, due to its complex structure. To address these problems, a PLS-based pruning algorithm is hereby proposed for a simplified LSTM (PSLSTM). First, a hybrid strategy is designed to simplify the internal structure of LSTM, which combines the structure simplification and parameter reduction for gates. Second, partial least squares (PLS) regression coefficients are used as the metric to evaluate the importance of the memory blocks, and the redundant hidden layer size is pruned by merging unimportant blocks with their most correlated ones. The Backpropagation Through Time (BPTT) algorithm is utilized as the learning algorithm to update the network parameters. Finally, several benchmark and practical datasets for time series prediction are used to evaluate the performance of the proposed PSLSTM. The experimental results demonstrate that the PLS-based pruning algorithm can achieve the trade-off between a good generalization ability and a compact network structure. The computational complexity is improved by the simple internal structure as well as the compact hidden layer size, without sacrificing prediction accuracy. (C) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页数:13
相关论文
共 52 条
[1]   Improving time series forecasting using LSTM and attention models [J].
Abbasimehr, Hossein ;
Paki, Reza .
JOURNAL OF AMBIENT INTELLIGENCE AND HUMANIZED COMPUTING, 2022, 13 (01) :673-691
[2]   An optimized model using LSTM network for demand forecasting [J].
Abbasimehr, Hossein ;
Shabani, Mostafa ;
Yousefi, Mohsen .
COMPUTERS & INDUSTRIAL ENGINEERING, 2020, 143
[3]   Predictive cardiac motion modeling and correction with partial least squares regression [J].
Ablitt, NA ;
Gao, JX ;
Keegan, J ;
Stegger, L ;
Firmin, DN ;
Yang, GZ .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2004, 23 (10) :1315-1324
[4]   LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT [J].
BENGIO, Y ;
SIMARD, P ;
FRASCONI, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02) :157-166
[5]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[6]   Cooperative coevolution of Elman recurrent neural networks for chaotic time series prediction [J].
Chandra, Rohitash ;
Zhang, Mengjie .
NEUROCOMPUTING, 2012, 86 :116-123
[7]  
Chen H., 2019, VIA MANIFOLD OPTIMIZ, V30, P588
[8]   The optimal design and application of LSTM neural network based on the hybrid coding PSO algorithm [J].
Chen, Zhonglin ;
Yang, Cuili ;
Qiao, Junfei .
JOURNAL OF SUPERCOMPUTING, 2022, 78 (05) :7227-7259
[9]  
Cho K., 2014, COMPUT SCI
[10]  
Ergen T., 2017, IEEE T NEURAL NETW L, P1, DOI DOI 10.1109/TNNLS.2017.2741598