Discrepancy-Based Theory and Algorithms for Forecasting Non-Stationary Time Series

被引:24
作者
Kuznetsov, Vitaly [1 ]
Mohri, Mehryar [2 ]
机构
[1] Google Res, 76 Ninth Ave, New York, NY 10011 USA
[2] Google Res & Courant Inst, 251 Mercer St, New York, NY 10012 USA
关键词
Time series; Forecasting; Non-stationary; Non-mixing; Generalization bounds; Discrepancy; Expected sequential covering numbers; Sequential Rademacher complexity; CONVERGENCE; PREDICTION; BOUNDS;
D O I
10.1007/s10472-019-09683-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present data-dependent learning bounds for the general scenario of non-stationary non-mixing stochastic processes. Our learning guarantees are expressed in terms of a data-dependent measure of sequential complexity and a discrepancy measure that can be estimated from data under some mild assumptions. Our learning bounds guide the design of new algorithms for non-stationary time series forecasting for which we report several favorable experimental results.
引用
收藏
页码:367 / 399
页数:33
相关论文
共 45 条
[1]   UNIFORM CONVERGENCE OF VAPNIK-CHERVONENKIS CLASSES UNDER ERGODIC SAMPLING [J].
Adams, Terrence M. ;
Nobel, Andrew B. .
ANNALS OF PROBABILITY, 2010, 38 (04) :1345-1367
[2]   The Generalization Ability of Online Algorithms for Dependent Data [J].
Agarwal, Alekh ;
Duchi, John C. .
IEEE TRANSACTIONS ON INFORMATION THEORY, 2013, 59 (01) :573-587
[3]  
Andrews D.:, 1983, 664 YAL U COWL FDN R, V664
[4]  
[Anonymous], 2014, Dependence Modelling
[5]  
[Anonymous], NIPS
[6]  
[Anonymous], 2017, P INT C LEARN REPR
[7]  
[Anonymous], MACHINE LEARNING
[8]  
[Anonymous], AISTAT
[9]  
[Anonymous], NIPS
[10]  
[Anonymous], 2015, PROBABILITY THEORY R