Efficient Implementations of Echo State Network Cross-Validation

被引:18
作者
Lukosevicius, Mantas [1 ]
Uselis, Arnas [1 ]
机构
[1] Kaunas Univ Technol, Studentu St 50-406, LT-51368 Kaunas, Lithuania
关键词
Echo state networks; Reservoir computing; Recurrent neural networks; Cross-validation; Time complexity;
D O I
10.1007/s12559-021-09849-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-Validation (CV) is still uncommon in time series modeling. Echo State Networks (ESNs), as a prime example of Reservoir Computing (RC) models, are known for their fast and precise one-shot learning, that often benefit from good hyper-parameter tuning. This makes them ideal to change the status quo. We discuss CV of time series for predicting a concrete time interval of interest, suggest several schemes for cross-validating ESNs and introduce an efficient algorithm for implementing them. This algorithm is presented as two levels of optimizations of doing k-fold CV. Training an RC model typically consists of two stages: (i) running the reservoir with the data and (ii) computing the optimal readouts. The first level of our optimization addresses the most computationally expensive part (i) and makes it remain constant irrespective of k. It dramatically reduces reservoir computations in any type of RC system and is enough if k is small. The second level of optimization also makes the (ii) part remain constant irrespective of large k, as long as the dimension of the output is low. We discuss when the proposed validation schemes for ESNs could be beneficial, three options for producing the final model and empirically investigate them on six different real-world datasets, as well as do empirical computation time experiments. We provide the code in an online repository. Proposed CV schemes give better and more stable test performance in all the six different real-world datasets, three task types. Empirical run times confirm our complexity analysis. In most situations, k-fold CV of ESNs and many other RC models can be done for virtually the same time and space complexity as a simple single-split validation. This enables CV to become a standard practice in RC.
引用
收藏
页码:1470 / 1484
页数:15
相关论文
共 26 条
  • [1] [Anonymous], 2007, Scholarpedia, DOI 10.4249/scholarpedia.2330
  • [2] A survey of cross-validation procedures for model selection
    Arlot, Sylvain
    Celisse, Alain
    [J]. STATISTICS SURVEYS, 2010, 4 : 40 - 79
  • [3] LEARNING LONG-TERM DEPENDENCIES WITH GRADIENT DESCENT IS DIFFICULT
    BENGIO, Y
    SIMARD, P
    FRASCONI, P
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (02): : 157 - 166
  • [4] A note on the validity of cross-validation for evaluating autoregressive time series prediction
    Bergmeir, Christoph
    Hyndman, Rob J.
    Koo, Bonsoo
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2018, 120 : 70 - 83
  • [5] Evaluating time series forecasting models: an empirical study on performance estimation methods
    Cerqueira, Vitor
    Torgo, Luis
    Mozetic, Igor
    [J]. MACHINE LEARNING, 2020, 109 (11) : 1997 - 2028
  • [6] Daukantas, 2010, P 14 INT C BIOMEDICA, P135
  • [7] STATISTICAL-THEORY - THE PREQUENTIAL APPROACH
    DAWID, AP
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES A-STATISTICS IN SOCIETY, 1984, 147 : 278 - 292
  • [8] DOYA K, 1992, 1992 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOLS 1-6, P2777, DOI 10.1109/ISCAS.1992.230622
  • [9] Harnessing nonlinearity: Predicting chaotic systems and saving energy in wireless communication
    Jaeger, H
    Haas, H
    [J]. SCIENCE, 2004, 304 (5667) : 78 - 80
  • [10] Jaeger H., 2003, NIPS 2002 P 15 INT C, P609