A hybrid framework for sequential data prediction with end-to-end optimization

被引:0
作者
Aydin, Mustafa E. [1 ]
Kozat, Suleyman S. [1 ]
机构
[1] Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkey
关键词
Feature extraction; End-to-end learning; Online learning; Prediction; Long short-term memory (LSTM); Soft gradient boosting decision tree (sGBDT);
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We investigate nonlinear prediction in an online setting and introduce a hybrid model that effectively mitigates, via an end-to-end architecture, the need for hand-designed features and manual model selection issues of conventional nonlinear prediction/regression methods. In particular, we use recursive structures to extract features from sequential signals, while preserving the state information, i.e., the history, and boosted decision trees to produce the final output. The connection is in an end-to-end fashion and we jointly optimize the whole architecture using stochastic gradient descent, for which we also provide the backward pass update equations. In particular, we employ a recurrent neural network (LSTM) for adaptive feature extraction from sequential data and a gradient boosting machinery (soft GBDT) for effective supervised regression. Our framework is generic so that one can use other deep learning architectures for feature extraction (such as RNNs and GRUs) and machine learning algorithms for decision making as long as they are differentiable. We demonstrate the learning behavior of our algorithm on synthetic data and the significant performance improvements over the conventional methods over various real life datasets. Furthermore, we openly share the source code of the proposed method to facilitate further research. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页数:12
相关论文
共 49 条
  • [1] Alcala-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
  • [2] Alcoa Inc, COMM STOCK
  • [3] Distributed online gradient boosting on data stream over multi-agent networks
    An, Xibin
    Hu, Chen
    Liu, Gang
    Lin, Haoshen
    [J]. SIGNAL PROCESSING, 2021, 189
  • [4] Biau G., 2016, ARXIV
  • [5] Breiman L, 2017, CLASSIFICATION REGRE, DOI [10.2307/2530946, DOI 10.1201/9781315139470]
  • [6] XGBoost: A Scalable Tree Boosting System
    Chen, Tianqi
    Guestrin, Carlos
    [J]. KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, : 785 - 794
  • [7] Cho K., 2014, P 2014 C EMP METH NA
  • [8] Highly efficient hierarchical online nonlinear regression using second order methods
    Civek, Burak C.
    Delibalta, Ibrahim
    Kozat, Suleyman S.
    [J]. SIGNAL PROCESSING, 2017, 137 : 22 - 32
  • [9] THE TERNARY DECISION TREE - A MULTISTAGE CLASSIFIER WITH REJECT
    CRIGNON, I
    DUBUISSON, B
    LEGUILLOU, G
    [J]. SIGNAL PROCESSING, 1983, 5 (05) : 433 - 443
  • [10] Deng J, 2009, 2009 IEEE C COMP VIS, P248, DOI DOI 10.1109/CVPR.2009.5206848