Naive Bayes classifiers that perform well with continuous variables

被引:0
|
作者
Bouckaert, RR [1 ]
机构
[1] Univ Waikato, Dept Comp Sci, Dunedin, New Zealand
来源
AI 2004: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS | 2004年 / 3339卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are three main methods for handling continuous variables in naive Bayes classifiers, namely, the normal method (parametric approach), the kernel method (non parametric approach) and discretization. In this article, we perform a methodologically sound comparison of the three methods, which shows large mutual differences of each of the methods and no single method being universally better. This suggests that a method for selecting one of the three approaches to continuous variables could improve overall performance of the naive Bayes classifier. We present three methods that can be implemented efficiently v-fold cross validation for the normal, kernel and discretization method. Empirical evidence suggests that selection using 10 fold cross validation (especially when repeated 10 times) can largely and significantly improve over all performance of naive Bayes classifiers and consistently outperform any of the three popular methods for dealing with continuous variables on their own. This is remarkable, since selection among more classifiers does not consistently result in better accuracy.
引用
收藏
页码:1089 / 1094
页数:6
相关论文
共 50 条
  • [41] Exponential Loss Minimization for Learning Weighted Naive Bayes Classifiers
    Kim, Taeheung
    Lee, Jong-Seok
    IEEE ACCESS, 2022, 10 : 22724 - 22736
  • [42] On decision boundaries of naive Bayes in continuous domains
    Elomaa, T
    Rousu, J
    KNOWLEDGE DISCOVERY IN DATABASES: PKDD 2003, PROCEEDINGS, 2003, 2838 : 144 - 155
  • [43] The learning and optimization of full Bayes classifiers with continuous attributes
    Wang, S.-C. (wangsc@lixin.edu.cn), 1600, Science Press (35): : 2129 - 2138
  • [44] Risk based Government Audit Planning using Naive Bayes Classifiers`
    Balaniuk, Remis
    Bessiere, Pierre
    Mazer, Emmanuel
    Cobbe, Paulo
    ADVANCES IN KNOWLEDGE-BASED AND INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, 2012, 243 : 1313 - 1323
  • [45] Learning naive Bayes classifiers from positive and unlabelled examples with uncertainty
    He, Jiazhen
    Zhang, Yang
    Li, Xue
    Shi, Peng
    INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2012, 43 (10) : 1805 - 1825
  • [46] COMPARISON OF NAIVE BAYES AND SUPPORT VECTOR MACHINE CLASSIFIERS ON DOCUMENT CLASSIFICATION
    Moe, Zun Hlaing
    San, Thida
    Khin, Mie Mie
    Tin, Hlaing May
    2018 IEEE 7TH GLOBAL CONFERENCE ON CONSUMER ELECTRONICS (GCCE 2018), 2018, : 466 - 467
  • [47] A new decision to take for cost-sensitive Naive Bayes classifiers
    Di Nunzio, Giorgio Maria
    INFORMATION PROCESSING & MANAGEMENT, 2014, 50 (05) : 653 - 674
  • [48] Human activity classification using Decision Tree and Naive Bayes classifiers
    Maswadi, Kholoud
    Ghani, Norjihan Abdul
    Hamid, Suraya
    Rasheed, Muhammads Babar
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (14) : 21709 - 21726
  • [49] Efficient weighted naive bayes classifiers to predict air quality index
    Jasleen Kaur Sethi
    Mamta Mittal
    Earth Science Informatics, 2022, 15 : 541 - 552
  • [50] Classification for Authorship of Tweets by Comparing Logistic Regression and Naive Bayes Classifiers
    Aborisade, Opeyemi Mulikat
    Anwar, Mohd
    2018 IEEE INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION (IRI), 2018, : 269 - 276