Attention-Based SeriesNet: An Attention-Based Hybrid Neural Network Model for Conditional Time Series Forecasting

被引:4
作者
Cheng, Yepeng [1 ]
Liu, Zuren [1 ]
Morimoto, Yasuhiko [1 ]
机构
[1] Hiroshima Univ, Grad Sch Engn, Dept Informat Engn, Higashihiroshima 7398527, Japan
关键词
attention; convolutional neural network; recurrent neural network;
D O I
10.3390/info11060305
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Traditional time series forecasting techniques can not extract good enough sequence data features, and their accuracies are limited. The deep learning structure SeriesNet is an advanced method, which adopts hybrid neural networks, including dilated causal convolutional neural network (DC-CNN) and Long-short term memory recurrent neural network (LSTM-RNN), to learn multi-range and multi-level features from multi-conditional time series with higher accuracy. However, they didn't consider the attention mechanisms to learn temporal features. Besides, the conditioning method for CNN and RNN is not specific, and the number of parameters in each layer is tremendous. This paper proposes the conditioning method for two types of neural networks, and respectively uses the gated recurrent unit network (GRU) and the dilated depthwise separable temporal convolutional networks (DDSTCNs) instead of LSTM and DC-CNN for reducing the parameters. Furthermore, this paper presents the lightweight RNN-based hidden state attention module (HSAM) combined with the proposed CNN-based convolutional block attention module (CBAM) for time series forecasting. Experimental results show our model is superior to other models from the viewpoint of forecasting accuracy and computation efficiency.
引用
收藏
页数:15
相关论文
共 29 条
[1]  
Abien F.A., 2018, ARXIV180308375V2
[2]  
[Anonymous], 2012, On the Difficulty of Training Recurrent Neural Networks, DOI DOI 10.48550/ARXIV.1211.5063
[3]   Dilated convolutional neural networks for time series forecasting [J].
Borovykh, Anastasia ;
Bohte, Sander ;
Oosterlee, Cornelis W. .
JOURNAL OF COMPUTATIONAL FINANCE, 2019, 22 (04) :73-101
[4]  
Borovykh A, 2017, LECT NOTES COMPUT SC, V10614, P729
[5]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[6]  
Chung J., 2014, EMPIRICAL EVALUATION, P1
[7]   Position-Based Content Attention for Time Series Forecasting with Sequence-to-Sequence RNNs [J].
Cinar, Yagmur Gizem ;
Mirisaee, Hamid ;
Goswami, Parantapa ;
Gaussier, Eric ;
Ait-Bachir, Ali ;
Strijov, Vadim .
NEURAL INFORMATION PROCESSING, ICONIP 2017, PT V, 2017, 10638 :533-544
[8]  
Drucker H, 1997, ADV NEUR IN, V9, P155
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
Hochreiter S., 1997, Neural Computation, V9, P1735