Robust recurrent neural networks for time series forecasting

被引:45
作者
Zhang, Xueli [1 ]
Zhong, Cankun [1 ]
Zhang, Jianjun [1 ]
Wang, Ting [2 ]
Ng, Wing W. Y. [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangdong Prov Key Lab Computat Intelligence & Cyb, Guangzhou 510006, Peoples R China
[2] South China Univ Technol, Guangzhou Peoples Hosp 1, Sch Med, Dept Radiol, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Recurrent neural networks (RNNs); Localized stochastic sensitivity (LSS); Time series forecasting; LSTM; CLASSIFICATION; PREDICTION; ERROR; MODEL;
D O I
10.1016/j.neucom.2023.01.037
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) are widely utilized in time series forecasting tasks. In practical appli-cations, there are noises in real-life time series data. A model's generalization capacity will be diminished by model uncertainty with regard to input noises. However, the robustness of RNNs with respect to input noises has not been well studied yet. The localized stochastic sensitivity (LSS), which measures output disturbances with respect to input perturbations of learning models, has been successfully applied to improve the robustness of different neural networks on tabular and image data. But its effectiveness on time series data has not been explored. Therefore, we extend the idea of LSS and apply it to the robust RNNs training for time series forecasting problems. With the minimization of LSS, output sensitivities of RNNs with respect to small perturbations are reduced. So, the proposed robust RNNs will not be affected by slight input noises. We have used the LSTM as an example for theoretical analysis and analyzed the effectiveness of LSS on several RNN variants including the vanilla RNN, the gated recurrent unit (GRU), the long-short-term memory (LSTM), and the bi-directional long short-term memory (Bi-LSTM) in empir-ical studies. Experimental results confirm the efficiency of applying LSS to enhance the robustness of RNNs for time series data. For example, the robust RNNs outperform their counterpart by 53.26% and 49.45% on average in terms of Root Mean Squared Error and R-Square on five datasets.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页码:143 / 157
页数:15
相关论文
共 46 条
[1]   Describing Video With Attention-Based Bidirectional LSTM [J].
Bin, Yi ;
Yang, Yang ;
Shen, Fumin ;
Xie, Ning ;
Shen, Heng Tao ;
Li, Xuelong .
IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (07) :2631-2641
[2]  
Britz D, 2017, Arxiv, DOI [arXiv:1703.03906, DOI 10.48550/ARXIV.1703.03906, 10.48550/arXiv.1703.03906]
[3]   Fractional Neuro-Sequential ARFIMA-LSTM for Financial Market Forecasting [J].
Bukhari, Ayaz Hussain ;
Raja, Muhammad Asif Zahoor ;
Sulaiman, Muhammad ;
Islam, Saeed ;
Shoaib, Muhammad ;
Kumam, Poom .
IEEE ACCESS, 2020, 8 :71326-71338
[4]  
Lipton ZC, 2015, Arxiv, DOI [arXiv:1506.00019, DOI 10.48550/ARXIV.1506.00019, 10.48550/ARXIV.1506.00019]
[5]  
Cho KYHY, 2014, Arxiv, DOI arXiv:1406.1078
[6]   A hybrid evolutionary decomposition system for time series forecasting [J].
de Oliveira, Joao F. L. ;
Ludermir, Teresa B. .
NEUROCOMPUTING, 2016, 180 :27-34
[7]   Learning precise timing with LSTM recurrent networks [J].
Gers, FA ;
Schraudolph, NN ;
Schmidhuber, J .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (01) :115-143
[8]  
Gers FA, 1999, IEE CONF PUBL, P850, DOI [10.1162/089976600300015015, 10.1049/cp:19991218]
[9]   Framewise phoneme classification with bidirectional LSTM and other neural network architectures [J].
Graves, A ;
Schmidhuber, J .
NEURAL NETWORKS, 2005, 18 (5-6) :602-610
[10]  
Graves A, 2007, LECT NOTES COMPUT SC, V4668, P549