Blending gradient boosted trees and neural networks for point and probabilistic forecasting of hierarchical time series

被引:5
作者
Nasios, Ioannis [1 ]
Vogklis, Konstantinos [1 ,2 ]
机构
[1] Nodalpoint Syst, Athens, Greece
[2] Ionian Univ, Dept Tourism, Corfu, Greece
关键词
M5; Competition; Point forecast; Probabilistic forecast; Regression models; Gradient boosted trees; Neural networks; Machine learning;
D O I
10.1016/j.ijforecast.2022.01.001
中图分类号
F [经济];
学科分类号
02 ;
摘要
In this study, we addressed the problem of point and probabilistic forecasting by describing a blending methodology for machine learning models from the gradient boosted trees and neural networks families. These principles were successfully applied in the recent M5 Competition in both the Accuracy and Uncertainty tracks. The key points of our methodology are: (a) transforming the task into regression on sales for a single day; (b) information-rich feature engineering; (c) creating a diverse set of state-of-the-art machine learning models; and (d) carefully constructing validation sets for model tuning. We show that the diversity of the machine learning models and careful selection of validation examples are most important for the effectiveness of our approach. Forecasting data have an inherent hierarchical structure (12 levels) but none of our proposed solutions exploited the hierarchical scheme. Using the proposed methodology, we ranked within the gold medal range in the Accuracy track and within the prizes in the Uncertainty track. Inference code with pre-trained models are available on GitHub.1 (c) 2022 International Institute of Forecasters. Published by Elsevier B.V. All rights reserved.
引用
收藏
页码:1448 / 1459
页数:12
相关论文
共 24 条
  • [1] Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
  • [2] Akbik A, 2019, NAACL HLT 2019: THE 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES: PROCEEDINGS OF THE DEMONSTRATIONS SESSION, P54
  • [3] Al-Rfou R., 2013, PREPRINT
  • [4] Ben Taieb Souhaib, 2012, Recursive and direct multi-step forecasting: the best of both worlds, V19
  • [5] Bengio Y, 2004, J MACH LEARN RES, V5, P1089
  • [6] Bishop C.M., 1995, NEURAL NETWORKS PATT
  • [7] Chollet F., 2018, Deep Learning with Python, V361
  • [8] Advances in forecasting with neural networks? Empirical evidence from the NN3 competition on time series prediction
    Crone, Sven F.
    Hibon, Michele
    Nikolopoulos, Konstantinos
    [J]. INTERNATIONAL JOURNAL OF FORECASTING, 2011, 27 (03) : 635 - 660
  • [9] Greedy function approximation: A gradient boosting machine
    Friedman, JH
    [J]. ANNALS OF STATISTICS, 2001, 29 (05) : 1189 - 1232
  • [10] Stochastic gradient boosting
    Friedman, JH
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2002, 38 (04) : 367 - 378