Application of Random Forest Algorithm on Feature Subset Selection and Classification and Regression

被引:120
作者
Jaiswal, Jitendra Kumar [1 ]
Samikannu, Rita [1 ]
机构
[1] VIT Univ, Sch Adv Sci, Vellore, Tamil Nadu, India
来源
2017 2ND WORLD CONGRESS ON COMPUTING AND COMMUNICATION TECHNOLOGIES (WCCCT) | 2017年
关键词
Random Forest; Feature Selection; Classification; Regression; Gini Index; Wrapper and Filter Methods;
D O I
10.1109/WCCCT.2016.25
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Feature subset selection becomes quite important and predominant in the case of data sets those are contained with higher number of variables. It discards insignificant variables and produces efficient and improved prediction performance on the class variables that is more cost effective and more reliable understanding of the data. Random forest has been emerged as a quite efficient and robust algorithm that can handle feature selection problem even with the higher number of variables. It is also very much efficient while dealing with Missing data imputation, classification, and regression problems. It can also handle outliers and noisy data very well. In this paper we applied the concept of random forest algorithm on the feature subset selection and classification and regression to perform the comparative study of the random forest algorithm in different perspectives.
引用
收藏
页码:65 / 68
页数:4
相关论文
共 7 条
[1]  
Bekkerman R., 2003, Journal of Machine Learning Research, V3, P1183, DOI 10.1162/153244303322753625
[2]   Selection of relevant features and examples in machine learning [J].
Blum, AL ;
Langley, P .
ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) :245-271
[3]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[4]  
Caruana R., 2003, Journal of Machine Learning Research, V3, P1245, DOI 10.1162/153244303322753652
[5]  
Olkin I., 2013, INTRO STAT LEARNING, V8th, P204
[6]  
Schapire RE, 1998, ANN STAT, V26, P1651
[7]  
Tibshirani R. Bias, 1996, TECHNICAL REPORT