A Heterogeneous Ensemble of Trees

被引:0
作者
Cheng, Wen Xin [1 ]
Katuwal, Rakesh [1 ]
Suganthan, P. N. [1 ]
Qiu, Xueheng [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
来源
2017 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI) | 2017年
关键词
classification; decision tree; heterogeneous ensemble; CLASSIFICATION; REGRESSION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decision Tree is a simple but popular machine learning algorithm. Although a single decision tree is not as accurate as other state-of-the-art classifiers, the performance can be significantly improved by combining the predictions of several decision trees i.e. by creating an ensemble of trees. In this paper, we study decision trees and their ensembles viz. Bagged Decision Trees, Random Forest, Extremely Randomized Trees, Rotation Forest, Gradient Boosted Trees and AdaBoosted Trees, and assess their performance on several UCI datasets. In addition, we propose a new ensemble method, Heterogeneous Ensemble of trees, and compare its performance with existing tree based classifiers. The heterogeneous ensemble is built with three different ensemble of trees (Random Forest, Rotation Forest, and Extremely Randomized Trees) with equal proportions to boost the diversity of the trees in the ensemble. A weightage scheme based on out-of-bag error is employed to combine the prediction of various trees for the final output prediction. Based on the experiments performed on several UCI datasets, the Heterogeneous Ensemble of trees obtains the best rank compared with other tree based classifiers.
引用
收藏
页码:1555 / 1560
页数:6
相关论文
共 23 条
  • [1] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [2] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [3] Breiman L, 1996, Tech. Rep. 460
  • [4] Criminisil A, 2011, FOUND TRENDS COMPUT, V7, P81, DOI [10.1561/0600000035, 10.1501/0000000035]
  • [5] Ensemble methods in machine learning
    Dietterich, TG
    [J]. MULTIPLE CLASSIFIER SYSTEMS, 2000, 1857 : 1 - 15
  • [6] Fernández-Delgado M, 2014, J MACH LEARN RES, V15, P3133
  • [7] Freund Y., 1996, P 13 INT C MACH LEAR, V96, P148, DOI DOI 10.5555/3091696.3091715
  • [8] Greedy function approximation: A gradient boosting machine
    Friedman, JH
    [J]. ANNALS OF STATISTICS, 2001, 29 (05) : 1189 - 1232
  • [9] On bias, variance, 0/1 - Loss, and the curse-of-dimensionality
    Friedman, JH
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 1997, 1 (01) : 55 - 77
  • [10] Gashler Mike, 2008, 2008 Seventh international conference on machine learning and applications, P900