A Random Forest approach using imprecise probabilities

被引:45
作者
Abellan, Joaquin [1 ]
Mantas, Carlos J. [1 ]
Castellano, Javier G. [1 ]
机构
[1] Univ Granada, Dept Comp Sci & Artificial Intelligence, Granada, Spain
关键词
Classification; Class noise; Random Forest; Imprecise probabilities; Uncertainty measures; DECISION TREES; CLASSIFICATION; ENTROPY; NOISE; ENSEMBLES;
D O I
10.1016/j.knosys.2017.07.019
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Random Forest classifier has been considered as an important reference in the data mining area. The building procedure of its base classifier (a decision tree) is principally based on a randomization process of data and features; and on a split criterion, which uses classic precise probabilities, to quantify the gain of information. One drawback found on this classifier is that it has a bad performance when it is applied on data sets with class noise. Very recently, it is proved that a new criterion which uses imprecise probabilities and general uncertainty measures, can improve the performance of the classic split criteria. In this work, the base classifier of the Random Forest is modified using that new criterion, producing also a new single decision tree model. This model join with the randomization process of features is the base classifier of a new procedure similar to the Random Forest, called Credal Random Forest. The principal differences between those two models are presented. In an experimental study, it is shown that the new method represents an improvement of the Random Forest when both are applied on data sets without class noise. But this improvement is notably greater when they are applied on data sets with class noise. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:72 / 84
页数:13
相关论文
共 43 条
[2]   Disaggregated total uncertainty measure for credal sets [J].
Abellán, J ;
Klir, GJ ;
Moral, S .
INTERNATIONAL JOURNAL OF GENERAL SYSTEMS, 2006, 35 (01) :29-44
[3]   Upper entropy of credal sets.: Applications to credal classification [J].
Abellán, J ;
Moral, S .
INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2005, 39 (2-3) :235-255
[4]   Building classification trees using the total uncertainty criterion [J].
Abellán, J ;
Moral, S .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2003, 18 (12) :1215-1225
[5]   Analyzing properties of Deng entropy in the theory of evidence [J].
Abellan, Joaquin .
CHAOS SOLITONS & FRACTALS, 2017, 95 :195-199
[8]   Bagging schemes on the presence of class noise in classification [J].
Abellan, Joaquin ;
Masegosa, Andres R. .
EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (08) :6827-6837
[9]   A FILTER-WRAPPER METHOD TO SELECT VARIABLES FOR THE NAIVE BAYES CLASSIFIER BASED ON CREDAL DECISION TREES [J].
Abellan, Joaquin ;
Masegosa, Andres R. .
INTERNATIONAL JOURNAL OF UNCERTAINTY FUZZINESS AND KNOWLEDGE-BASED SYSTEMS, 2009, 17 (06) :833-854
[10]  
Abellán J, 2009, LECT NOTES COMPUT SC, V5590, P446, DOI 10.1007/978-3-642-02906-6_39