A fuzzy random forest

被引:129
作者
Bonissone, Piero [2 ]
Cadenas, Jose M. [1 ]
Carmen Garrido, M. [1 ]
Andres Diaz-Valladares, R. [3 ]
机构
[1] Univ Murcia, Dept Ingn Informac & Comun, E-30001 Murcia, Spain
[2] GE Global Res, Moscow 123098, Russia
[3] Univ Montemorelos, Dept Ciencias Computac, Mexico City, DF, Mexico
关键词
Random forest; Fuzzy decision tree; Combination methods; Fuzzy sets; DECISION TREE; CLASSIFICATION; CLASSIFIERS; ENSEMBLES;
D O I
10.1016/j.ijar.2010.02.003
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
When individual classifiers are combined appropriately, a statistically significant increase in classification accuracy is usually obtained. Multiple classifier systems are the result of combining several individual classifiers. Following Breiman's methodology, in this paper a multiple classifier system based on a "forest" of fuzzy decision trees, i.e., a fuzzy random forest, is proposed. This approach combines the robustness of multiple classifier systems, the power of the randomness to increase the diversity of the trees, and the flexibility of fuzzy logic and fuzzy sets for imperfect data management. Various combination methods to obtain the final decision of the multiple classifier system are proposed and compared. Some of them are weighted combination methods which make a weighting of the decisions of the different elements of the multiple classifier system (leaves or trees). A comparative study with several datasets is made to show the efficiency of the proposed multiple classifier system and the various combination methods. The proposed multiple classifier system exhibits a good accuracy classification, comparable to that of the best classifiers when tested with conventional data sets. However, unlike other classifiers, the proposed classifier provides a similar accuracy when tested with imperfect datasets (with missing and fuzzy values) and with datasets with noise. (C) 2010 Elsevier Inc. All rights reserved.
引用
收藏
页码:729 / 747
页数:19
相关论文
共 32 条
[1]   Classification by ensembles from random partitions of high-dimensional data [J].
Ahn, Hongshik ;
Moon, Hojin ;
Fazzari, Melissa J. ;
Lim, Noha ;
Chen, James J. ;
Kodell, Ralph L. .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2007, 51 (12) :6166-6179
[2]  
Asuncion A., UCI MACHINE LEARNING
[3]   A comparison of decision tree ensemble creation techniques [J].
Banfield, Robert E. ;
Hall, Lawrence O. ;
Bowyer, Kevin W. ;
Kegelmeyer, W. P. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2007, 29 (01) :173-180
[4]   CONTROLLING THE FALSE DISCOVERY RATE - A PRACTICAL AND POWERFUL APPROACH TO MULTIPLE TESTING [J].
BENJAMINI, Y ;
HOCHBERG, Y .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1995, 57 (01) :289-300
[5]  
BONISSONE PP, 2008, APPL SOFT COMPUT, DOI DOI 10.1016/J.ASOC.2008.03.06
[6]  
BONISSONE PP, 2008, INF PROC MAN UNC KNO, P1231
[7]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[8]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[9]  
CADENAS JM, P 2 S LOG FUZZ SOFT, P41
[10]  
CADENAS JM, 2009, P 13 C AS ESP INT AR, P379