OPTIMAL PARTITIONING FOR CLASSIFICATION AND REGRESSION TREES

被引:153
作者
CHOU, PA [1 ]
机构
[1] STANFORD UNIV, DEPT ELECT ENGN, STANFORD, CA 94305 USA
关键词
CLUSTERING; DECISION TREES; INFORMATION DIVERGENCE; TEXT-TO-SPEECH;
D O I
10.1109/34.88569
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In designing a decision tree for classification or regression, one selects at each node a feature to be tested, and partitions the range of that feature into a small number of bins, each bin corresponding to a child of the node. When the feature's range is discrete with N unordered outcomes, the optimal partition, that is, the partition minimizing an expected loss, is usually found by an exhaustive search through all possible partitions. Since the number of possible partitions grows exponentially in N, this approach is impractical when N is larger than about 10 or 20. In this paper, we present an iterative algorithm that finds a locally optimal partition for an arbitrary loss function, in time linear in N for each iteration. The algorithm is a K-means like clustering algorithm that uses as its distance measure a generalization of Kullback's information divergence. Moreover, we prove that the globally optimal partition must satisfy a nearest neighbor condition using divergence as the distance measure. These results generalize similar results of Breiman et al. to an arbitrary number of classes or regression variables and to an arbitrary number of bins. We also provide experimental results on a text-to-speech example, and we suggest additional applications of the algorithm, including the design of variable combinations, surrogate splits, composite nodes, and decision graphs.
引用
收藏
页码:340 / 354
页数:15
相关论文
共 50 条
[41]   PAMC: Partitioning around Medoids for Classification [J].
Department of Computer Applications, Samrat Ashok Technological Institute, Vidisha -464001, India .
Inf. Technol. J., 2006, 6 (1102-1105) :1102-1105
[42]   Optimal partitioning for the proportional hazards model [J].
Govindarajulu, Usha ;
Tarpey, Thaddeus .
JOURNAL OF APPLIED STATISTICS, 2022, 49 (04) :968-987
[43]   Classification of Daily Body Weight Gains in Beef Calves Using Decision Trees, Artificial Neural Networks, and Logistic Regression [J].
Grzesiak, Wilhelm ;
Zaborski, Daniel ;
Pilarczyk, Renata ;
Wojcik, Jerzy ;
Adamczyk, Krzysztof .
ANIMALS, 2023, 13 (12)
[44]   Prediction of ordinal classes using regression trees [J].
Kramer, S ;
Pfahringer, B ;
Widmer, G ;
De Groeve, M .
FUNDAMENTA INFORMATICAE, 2001, 47 (1-2) :1-13
[45]   Elgasir: An algorithm for creating Fuzzy Regression Trees [J].
Gasir, Fathi ;
Bandar, Zuhair ;
Crockett, Keeley .
2009 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 2009, :332-337
[46]   Predictive modelling for solar thermal energy systems: A comparison of support vector regression, random forest, extra trees and regression trees [J].
Ahmad, Muhammad Waseem ;
Reynolds, Jonathan ;
Rezgui, Yacine .
JOURNAL OF CLEANER PRODUCTION, 2018, 203 :810-821
[47]   Modeling of wastewater treatment plant in Hama city using regression and regression trees [J].
Bodaka, Heba ;
Farhoud, Nahed ;
Hlali, Eyad .
ENVIRONMENTAL HEALTH ENGINEERING AND MANAGEMENT JOURNAL, 2023, 10 (03) :293-300
[48]   Expressive tests for classification and regression [J].
Morishita, S ;
Nakaya, A .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2000, E83D (01) :52-60
[49]   A Hybrid Approach Based on Decision Trees and Clustering for Breast Cancer Classification [J].
Elouedi, Hind ;
Meliani, Walid ;
Elouedi, Zied ;
Ben Amor, Nahla .
2014 6TH INTERNATIONAL CONFERENCE OF SOFT COMPUTING AND PATTERN RECOGNITION (SOCPAR), 2014, :226-231
[50]   Approximating Optimal Binary Decision Trees [J].
Micah Adler ;
Brent Heeringa .
Algorithmica, 2012, 62 :1112-1121