Oblique random forest ensemble via Least Square Estimation for time series forecasting

被引:76
作者
Qiu, Xueheng [1 ]
Zhang, Le [1 ]
Suganthan, Ponnuthurai Nagaratnam [1 ]
Amaratunga, Gehan A. J. [2 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, 50 Nanyang Ave, Singapore 639798, Singapore
[2] Univ Cambridge, Ctr Adv Photon & Elect, Elect Engn Div, Engn Dept, Cambridge CB3 0FA, England
基金
新加坡国家研究基金会;
关键词
Ensemble learning; Time series forecasting; Oblique random forest; Neural networks; Support vector regression; NEURAL-NETWORK; CLASSIFIERS; PREDICTION; DEEP; CLASSIFICATION; REGRESSION; ALGORITHM; MODEL;
D O I
10.1016/j.ins.2017.08.060
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent studies in Machine Learning indicates that the classifiers most likely to be the bests are the random forests. As an ensemble classifier, random forest combines multiple decision trees to significant decrease the overall variances. Conventional random forest employs orthogonal decision tree which selects one "optimal" feature to split the data instances within a non-leaf node according to impurity criteria such as Gini impurity, information gain and so on. However, orthogonal decision tree may fail to capture the geometrical structure of the data samples. Motivated by this, we make the first attempt to study the oblique random forest in the context of time series forecasting. In each node of the decision tree, instead of the single "optimal" feature based orthogonal classification algorithms used by standard random forest, a least square classifier is employed to perform partition. The proposed method is advantageous with respect to both efficiency and accuracy. We empirically evaluate the proposed method on eight generic time series datasets and five electricity load demand time series datasets from the Australian Energy Market Operator and compare with several other benchmark methods. (C) 2017 Elsevier Inc. All rights reserved.
引用
收藏
页码:249 / 262
页数:14
相关论文
共 49 条
[21]  
Haykin S., 1994, NEURAL NETWORKS COMP, DOI 10.1142/S0129065794000372
[22]  
Hinton, 2012, NEURAL NETWORKS TRIC, P599, DOI [DOI 10.1007/978-3-642-35289-8_32, 10.1007/978-3-642-35289-832]
[23]   Reducing the dimensionality of data with neural networks [J].
Hinton, G. E. ;
Salakhutdinov, R. R. .
SCIENCE, 2006, 313 (5786) :504-507
[24]   Training products of experts by minimizing contrastive divergence [J].
Hinton, GE .
NEURAL COMPUTATION, 2002, 14 (08) :1771-1800
[25]   A fast learning algorithm for deep belief nets [J].
Hinton, Geoffrey E. ;
Osindero, Simon ;
Teh, Yee-Whye .
NEURAL COMPUTATION, 2006, 18 (07) :1527-1554
[26]  
Ho TK, 1998, IEEE T PATTERN ANAL, V20, P832, DOI 10.1109/34.709601
[27]  
Hyun Ah Song, 2013, Neural Information Processing. 20th International Conference, ICONIP 2013. Proceedings: LNCS 8226, P466, DOI 10.1007/978-3-642-42054-2_58
[28]   Joint analysis of two microarray gene-expression data sets to select lung adenocarcinoma marker genes [J].
Jiang, HY ;
Deng, YP ;
Chen, HS ;
Tao, L ;
Sha, QY ;
Chen, J ;
Tsai, CJ ;
Zhang, SL .
BMC BIOINFORMATICS, 2004, 5 (1)
[29]   Investigating macro-level hotzone identification and variable importance using big data: A random forest models approach [J].
Jiang, Ximiao ;
Abdel-Aty, Mohamed ;
Hu, Jia ;
Lee, Jaeyoung .
NEUROCOMPUTING, 2016, 181 :53-63
[30]   Short-term load forecasting using SVR (support vector regression)-based radial basis function neural network with dual extended Kalman filter [J].
Ko, Chia-Nan ;
Lee, Cheng-Ming .
ENERGY, 2013, 49 :413-422