Naive Bayes classifiers that perform well with continuous variables

被引:0
|
作者
Bouckaert, RR [1 ]
机构
[1] Univ Waikato, Dept Comp Sci, Dunedin, New Zealand
来源
AI 2004: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS | 2004年 / 3339卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are three main methods for handling continuous variables in naive Bayes classifiers, namely, the normal method (parametric approach), the kernel method (non parametric approach) and discretization. In this article, we perform a methodologically sound comparison of the three methods, which shows large mutual differences of each of the methods and no single method being universally better. This suggests that a method for selecting one of the three approaches to continuous variables could improve overall performance of the naive Bayes classifier. We present three methods that can be implemented efficiently v-fold cross validation for the normal, kernel and discretization method. Empirical evidence suggests that selection using 10 fold cross validation (especially when repeated 10 times) can largely and significantly improve over all performance of naive Bayes classifiers and consistently outperform any of the three popular methods for dealing with continuous variables on their own. This is remarkable, since selection among more classifiers does not consistently result in better accuracy.
引用
收藏
页码:1089 / 1094
页数:6
相关论文
共 50 条
  • [21] On why discretization works for naive-Bayes classifiers
    Yang, Y
    Webb, GI
    AI 2003: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2003, 2903 : 440 - 452
  • [22] Naive Bayes classifiers learned from incomplete data
    Chen, Jingnian
    Huang, Houkuan
    Tian, Fengzhan
    Qiao, Zhufeng
    Jisuanji Gongcheng/Computer Engineering, 2006, 32 (17): : 86 - 88
  • [23] On the efficiency of data collection for multiple Naive Bayes classifiers
    Manino, Edoardo
    Long Tran-Thanh
    Jennings, Nicholas R.
    ARTIFICIAL INTELLIGENCE, 2019, 275 : 356 - 378
  • [24] Augmenting Naive Bayes Classifiers with Statistical Language Models
    Fuchun Peng
    Dale Schuurmans
    Shaojun Wang
    Information Retrieval, 2004, 7 : 317 - 345
  • [25] Maximum a posteriori tree augmented naive Bayes classifiers
    Cerquides, J
    de Màntaras, RL
    DISCOVERY SCIENCE, PROCEEDINGS, 2004, 3245 : 73 - 88
  • [26] A Novel Naive Bayes Voting Strategy for Combining Classifiers
    De Stefano, C.
    Fontanella, F.
    di Freca, A. Scotto
    13TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR 2012), 2012, : 467 - 472
  • [27] Incremental learning of tree augmented naive Bayes classifiers
    Alcobé, JR
    ADVANCES IN ARTIFICIAL INTELLIGENCE - IBERAMIA 2002, PROCEEDINGS, 2002, 2527 : 32 - 41
  • [28] Two feature weighting approaches for naive Bayes text classifiers
    Zhang, Lungan
    Jiang, Liangxiao
    Li, Chaoqun
    Kong, Ganggang
    KNOWLEDGE-BASED SYSTEMS, 2016, 100 : 137 - 144
  • [29] Aerial image texture classification based on naive Bayes classifiers
    Wuhan Daxue Xuebao (Xinxi Kexue Ban), 2006, 2 (108-111):
  • [30] Performance Comparison of New Fast Weighted Naive Bayes Classifier with Other Bayes Classifiers
    Aksoy, Gamzepelin
    Karabatak, Murat
    2019 7TH INTERNATIONAL SYMPOSIUM ON DIGITAL FORENSICS AND SECURITY (ISDFS), 2019,