A distance correlation-based approach to characterize the effectiveness of recurrent neural networks for time series forecasting

被引:0
作者
Salazar, Christopher [1 ]
Banerjee, Ashis G. [1 ,2 ]
机构
[1] Univ Washington, Dept Ind & Syst Engn, Seattle, WA 98195 USA
[2] Univ Washington, Dept Mech Engn, Seattle, WA 98195 USA
关键词
Recurrent neural network; Time series forecasting; Distance correlation; INFORMATION; DEPENDENCE;
D O I
10.1016/j.neucom.2025.129641
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Time series forecasting has received a lot of attention, with recurrent neural networks (RNNs) being one of the widely used models due to their ability to handle sequential data. Previous studies on RNN time series forecasting, however, show inconsistent outcomes and offer few explanations for performance variations among the datasets. In this paper, we provide an approach to link time series characteristics with RNN components via the versatile metric of distance correlation. This metric allows us to examine the information flow through the RNN activation layers to be able to interpret and explain their performance. We empirically show that the RNN activation layers learn the lag structures of time series well. However, they gradually lose this information over the span of a few consecutive layers, thereby worsening the forecast quality for series with large lag structures. We also show that the activation layers cannot adequately model moving average and heteroskedastic time series processes. Last, we generate heatmaps for visual comparisons of the activation layers for different choices of the network hyperparameters to identify which of them affect the forecast performance. Our findings can, therefore, aid practitioners in assessing the effectiveness of RNNs for given time series data without actually training and evaluating the networks.
引用
收藏
页数:15
相关论文
共 44 条
  • [1] Neural Network Layer Algebra: A Framework to Measure Capacity and Compression in Deep Learning
    Badias, Alberto
    Banerjee, Ashis G.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10380 - 10393
  • [2] Statistical Mechanics of Deep Learning
    Bahri, Yasaman
    Kadmon, Jonathan
    Pennington, Jeffrey
    Schoenholz, Sam S.
    Sohl-Dickstein, Jascha
    Ganguli, Surya
    [J]. ANNUAL REVIEW OF CONDENSED MATTER PHYSICS, VOL 11, 2020, 2020, 11 : 501 - 528
  • [3] LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT
    BENGIO, Y
    SIMARD, P
    FRASCONI, P
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02): : 157 - 166
  • [4] Bousqaoui H., 2021, International Journal of Electrical and Computer Engineering (IJECE), V11, P3319, DOI DOI 10.11591/IJECE.V11I4.PP3319-3328
  • [5] Brockwell Davis, 1996, Introduction to Time Series and Forecasting
  • [6] Chen MS, 2019, Arxiv, DOI arXiv:1910.12947
  • [7] An Updated Literature Review of Distance Correlation and Its Applications to Time Series
    Edelmann, Dominic
    Fokianos, Konstantinos
    Pitsillou, Maria
    [J]. INTERNATIONAL STATISTICAL REVIEW, 2019, 87 (02) : 237 - 262
  • [8] FINDING STRUCTURE IN TIME
    ELMAN, JL
    [J]. COGNITIVE SCIENCE, 1990, 14 (02) : 179 - 211
  • [9] Elsayed S, 2021, Arxiv, DOI arXiv:2101.02118
  • [10] Gao SY, 2015, JMLR WORKSH CONF PRO, V38, P277