Increasing Accuracy of Random Forest Algorithm by Decreasing Variance

被引:0
作者
Alshare, Somaya [1 ]
Abdullah, Malak [1 ]
Quwaider, Muhannad [1 ]
机构
[1] Jordan Univ Sci & Technol, Dept Comp Engn, Irbid, Jordan
来源
2022 13TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION SYSTEMS (ICICS) | 2022年
关键词
decision trees; ensemble learning; bias-variance tradeoff; bagging; Random Forest; CLASSIFICATION; MODELS; TREE;
D O I
10.1109/ICICS55353.2022.9811109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study aims to add a level of randomization to the process of building a tree within a random forest. This extra randomization step is achieved by adopting a sectioning technique of a feature's set of values to search for the optimal threshold at each tree node split. According to the proposed section-based random forest algorithm (SBRF), on each node split of the decision tree, the following steps are performed: first, sorting the chosen feature's values, then dividing them into equal sections; next, randomly pick a candidate threshold from each section, evaluate each candidate threshold against a predetermined criterion, and finally, choose the best candidate threshold among them. As a result, SBRF produces models of less variance and not higher bias than the models created by random forest, consequently decreasing the generalization error.
引用
收藏
页码:232 / 238
页数:7
相关论文
共 39 条
  • [1] E-learningDJUST: E-learning dataset from Jordan university of science and technology toward investigating the impact of COVID-19 pandemic on education
    Abdullah, Malak
    Al-Ayyoub, Mahmoud
    AlRawashdeh, Saif
    Shatnawi, Farah
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (16) : 11481 - 11495
  • [2] Abedalla Ayat, 2021, PeerJ Comput Sci, V7, pe607, DOI 10.7717/peerj-cs.607
  • [3] Increasing diversity in random forest learning algorithm via imprecise probabilities
    Abellan, Joaquin
    Mantas, Carlos J.
    Castellano, Javier G.
    Moral-Garcia, SerafIn
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2018, 97 : 228 - 243
  • [4] A bi-objective optimization method to produce a near-optimal number of classifiers and increase diversity in Bagging
    Asadi, Shahrokh
    Roshan, Seyed Ehsan
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 213
  • [5] An empirical comparison of voting classification algorithms: Bagging, boosting, and variants
    Bauer, E
    Kohavi, R
    [J]. MACHINE LEARNING, 1999, 36 (1-2) : 105 - 139
  • [6] Reconciling modern machine-learning practice and the classical bias-variance trade-off
    Belkin, Mikhail
    Hsu, Daniel
    Ma, Siyuan
    Mandal, Soumik
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (32) : 15849 - 15854
  • [7] Regularization in statistics
    Bickel, Peter J.
    Li, Bo
    [J]. TEST, 2006, 15 (02) : 271 - 303
  • [8] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [9] Breiman L, 1996, MACH LEARN, V24, P123, DOI 10.1023/A:1018054314350
  • [10] Brown G., 2005, Information Fusion, V6, P5, DOI 10.1016/j.inffus.2004.04.004