A Review and Experimental Comparison of Multivariate Decision Trees

被引:21
作者
Canete-Sifuentes, Leonardo [1 ]
Monroy, Raul [1 ]
Medina-Perez, Miguel Angel [1 ]
机构
[1] Tecnol Monterrey, Sch Engn & Sci, Atizapan De Zaragoza 52926, Estado De Mexic, Mexico
关键词
Databases; Decision trees; Taxonomy; Feature extraction; Vegetation; Iterative algorithms; Indexes; Supervised classification; decision trees; multivariate decision trees; machine learning; CLASSIFICATION TREES; IMAGE; SELECTION; DATASET; SYSTEM; MODEL;
D O I
10.1109/ACCESS.2021.3102239
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Decision trees are popular as stand-alone classifiers or as base learners in ensemble classifiers. Mostly, this is due to decision trees having the advantage of being easy to explain. To improve the classification performance of decision trees, some authors have used Multivariate Decision Trees (MDTs), which allow combinations of features when splitting a node. While there is growing interest in the area, recent research in MDTs all have in common that they do not provide adequate comparison of related work: they do not consider relevant rival techniques, or they test algorithm performance in an insufficient number of databases. As a result, claims have no statistical sustain and, hence, there is a lack of general understanding of the actual capabilities of existing MDT induction algorithms, crucial to improving the state-of-the-art. In this paper, we report on an exhaustive review of MDTs. In particular, we give an overview of 37 MDT induction algorithms, out of which we have experimentally compared 19 of them in 57 databases. We provide a statistical comparison in all databases and subsets of databases according to the number of classes, number of features, number of instances, and degree of class imbalance. This allows us to identify groups of top-performing algorithms for different types of databases.
引用
收藏
页码:110451 / 110479
页数:29
相关论文
共 82 条
[1]   Cline: A new decision-tree family [J].
Amasyali, M. Fatih ;
Ersoy, Okan .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (02) :356-363
[2]  
[Anonymous], 1984, CLASSICATION REGRESS
[3]  
[Anonymous], 2021, IEEE Trans. Broadcast.
[4]  
[Anonymous], 1997, P 14 INT C MACH LEAR
[5]  
[Anonymous], 2004, Advances in Neural Information Processing Systems
[6]   DTreeSim: A new approach to compute decision tree similarity using re-mining [J].
Bakirli, Gozde ;
Birant, Derya .
TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2017, 25 (01) :108-125
[7]  
Benavoli A, 2017, J MACH LEARN RES, V18
[8]   Enlarging the margins in perceptron decision trees [J].
Bennett, KP ;
Cristianini, N ;
Shawe-Taylor, J ;
Wu, DH .
MACHINE LEARNING, 2000, 41 (03) :295-313
[9]   Optimal classification trees [J].
Bertsimas, Dimitris ;
Dunn, Jack .
MACHINE LEARNING, 2017, 106 (07) :1039-1082
[10]   Classification by vertical and cutting multi-hyperplane decision tree induction [J].
Better, Marco ;
Glover, Fred ;
Samorani, Michele .
DECISION SUPPORT SYSTEMS, 2010, 48 (03) :430-436