Robust recurrent neural networks for time series forecasting

被引:45
作者
Zhang, Xueli [1 ]
Zhong, Cankun [1 ]
Zhang, Jianjun [1 ]
Wang, Ting [2 ]
Ng, Wing W. Y. [1 ]
机构
[1] South China Univ Technol, Sch Comp Sci & Engn, Guangdong Prov Key Lab Computat Intelligence & Cyb, Guangzhou 510006, Peoples R China
[2] South China Univ Technol, Guangzhou Peoples Hosp 1, Sch Med, Dept Radiol, Guangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Recurrent neural networks (RNNs); Localized stochastic sensitivity (LSS); Time series forecasting; LSTM; CLASSIFICATION; PREDICTION; ERROR; MODEL;
D O I
10.1016/j.neucom.2023.01.037
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) are widely utilized in time series forecasting tasks. In practical appli-cations, there are noises in real-life time series data. A model's generalization capacity will be diminished by model uncertainty with regard to input noises. However, the robustness of RNNs with respect to input noises has not been well studied yet. The localized stochastic sensitivity (LSS), which measures output disturbances with respect to input perturbations of learning models, has been successfully applied to improve the robustness of different neural networks on tabular and image data. But its effectiveness on time series data has not been explored. Therefore, we extend the idea of LSS and apply it to the robust RNNs training for time series forecasting problems. With the minimization of LSS, output sensitivities of RNNs with respect to small perturbations are reduced. So, the proposed robust RNNs will not be affected by slight input noises. We have used the LSTM as an example for theoretical analysis and analyzed the effectiveness of LSS on several RNN variants including the vanilla RNN, the gated recurrent unit (GRU), the long-short-term memory (LSTM), and the bi-directional long short-term memory (Bi-LSTM) in empir-ical studies. Experimental results confirm the efficiency of applying LSS to enhance the robustness of RNNs for time series data. For example, the robust RNNs outperform their counterpart by 53.26% and 49.45% on average in terms of Root Mean Squared Error and R-Square on five datasets.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页码:143 / 157
页数:15
相关论文
共 46 条
[21]   Simplified long short-term memory model for robust and fast prediction [J].
Liu, Yong ;
Hao, Xin ;
Zhang, Biling ;
Zhang, Yuyan .
PATTERN RECOGNITION LETTERS, 2020, 136 (136) :81-86
[22]   A temporal-aware LSTM enhanced by loss-switch mechanism for traffic flow forecasting [J].
Lu, Huakang ;
Ge, Zuhao ;
Song, Youyi ;
Jiang, Dazhi ;
Zhou, Teng ;
Qin, Jing .
NEUROCOMPUTING, 2021, 427 (427) :169-178
[23]   Stochastic Sensitivity Tree Boosting for Imbalanced Prediction Problems of Protein-Ligand Interaction Sites [J].
Ng, Wing W. Y. ;
Zhang, Yuda ;
Zhang, Jianjun ;
Wang, Debby D. ;
Wang, Fu Lee .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2021, 5 (03) :373-383
[24]  
Peng Yusheng, 2021, NEUROCOMPUTING
[25]   Bidirectional recurrent neural networks [J].
Schuster, M ;
Paliwal, KK .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1997, 45 (11) :2673-2681
[26]  
Shi XJ, 2015, ADV NEUR IN, V28
[27]   Host-Parasite: Graph LSTM-in-LSTM for Group Activity Recognition [J].
Shu, Xiangbo ;
Zhang, Liyan ;
Sun, Yunlian ;
Tang, Jinhui .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (02) :663-674
[28]   Unsupervised Multi-modal Neural Machine Translation [J].
Su, Yuanhang ;
Fan, Kai ;
Nguyen Bach ;
Kuo, C-C Jay ;
Huang, Fei .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :10474-10483
[29]   On extended long short-term memory and dependent bidirectional recurrent neural network [J].
Su, Yuanhang ;
Kuo, C-C Jay .
NEUROCOMPUTING, 2019, 356 :151-161
[30]   Deep sequential fusion LSTM network for image description [J].
Tang, Pengjie ;
Wang, Hanli ;
Kwong, Sam .
NEUROCOMPUTING, 2018, 312 :154-164