Naive Bayes classifiers that perform well with continuous variables

被引:0
|
作者
Bouckaert, RR [1 ]
机构
[1] Univ Waikato, Dept Comp Sci, Dunedin, New Zealand
来源
AI 2004: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS | 2004年 / 3339卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are three main methods for handling continuous variables in naive Bayes classifiers, namely, the normal method (parametric approach), the kernel method (non parametric approach) and discretization. In this article, we perform a methodologically sound comparison of the three methods, which shows large mutual differences of each of the methods and no single method being universally better. This suggests that a method for selecting one of the three approaches to continuous variables could improve overall performance of the naive Bayes classifier. We present three methods that can be implemented efficiently v-fold cross validation for the normal, kernel and discretization method. Empirical evidence suggests that selection using 10 fold cross validation (especially when repeated 10 times) can largely and significantly improve over all performance of naive Bayes classifiers and consistently outperform any of the three popular methods for dealing with continuous variables on their own. This is remarkable, since selection among more classifiers does not consistently result in better accuracy.
引用
收藏
页码:1089 / 1094
页数:6
相关论文
共 50 条
  • [31] Naive Bayes Classifiers for Music Emotion Classification Based on Lyrics
    An, Yunjing
    Sun, Shutao
    Wang, Shujuan
    2017 16TH IEEE/ACIS INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2017), 2017, : 635 - 638
  • [32] A New Feature Selection Approach to Naive Bayes Text Classifiers
    Zhang, Lungan
    Jiang, Liangxiao
    Li, Chaoqun
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2016, 30 (02)
  • [33] Handling Different Levels of Granularity within Naive Bayes Classifiers
    Ince, Kemal
    Klawonn, Frank
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2013, 2013, 8206 : 521 - 528
  • [34] Compression-based averaging of selective naive Bayes classifiers
    Boulle, Marc
    JOURNAL OF MACHINE LEARNING RESEARCH, 2007, 8 : 1659 - 1685
  • [35] A New Instance-weighting Naive Bayes Text Classifiers
    Wu, Yongcheng
    2018 IEEE INTERNATIONAL CONFERENCE OF INTELLIGENT ROBOTICS AND CONTROL ENGINEERING (IRCE), 2018, : 198 - 202
  • [36] The mysterious optimality of Naive Bayes: Estimation of the probability in the system of "classifiers"
    Kupervasser O.
    Pattern Recognition and Image Analysis, 2014, 24 (1) : 1 - 10
  • [37] Multiple Naive Bayes Classifiers Ensemble for Traffic Incident Detection
    Liu, Qingchao
    Lu, Jian
    Chen, Shuyan
    Zhao, Kangjia
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2014, 2014
  • [38] Privacy-preserving Training Algorithm for Naive Bayes Classifiers
    Wang, Rui
    Tang, Xiangyun
    Shen, Meng
    Zhu, Liehuang
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 5639 - 5644
  • [39] Naive Bayes text classifiers: a locally weighted learning approach
    Jiang, Liangxiao
    Cai, Zhihua
    Zhang, Harry
    Wang, Dianhong
    JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2013, 25 (02) : 273 - 286
  • [40] Combining active learning and boosting for Naive Bayes text classifiers
    Kim, HJ
    Kim, J
    ADVANCES IN WEB-AGE INFORMATION MANAGEMENT: PROCEEDINGS, 2004, 3129 : 519 - 527