Accurate parameter estimation for Bayesian network classifiers using hierarchical Dirichlet processes

被引:0
|
作者
François Petitjean
Wray Buntine
Geoffrey I. Webb
Nayyar Zaidi
机构
[1] Monash University,Faculty of Information Technology
来源
Machine Learning | 2018年 / 107卷
关键词
Bayesian network; Parameter estimation; Graphical models; Dirichlet processes; Smoothing; Classification;
D O I
暂无
中图分类号
学科分类号
摘要
This paper introduces a novel parameter estimation method for the probability tables of Bayesian network classifiers (BNCs), using hierarchical Dirichlet processes (HDPs). The main result of this paper is to show that improved parameter estimation allows BNCs to outperform leading learning methods such as random forest for both 0–1 loss and RMSE, albeit just on categorical datasets. As data assets become larger, entering the hyped world of “big”, efficient accurate classification requires three main elements: (1) classifiers with low-bias that can capture the fine-detail of large datasets (2) out-of-core learners that can learn from data without having to hold it all in main memory and (3) models that can classify new data very efficiently. The latest BNCs satisfy these requirements. Their bias can be controlled easily by increasing the number of parents of the nodes in the graph. Their structure can be learned out of core with a limited number of passes over the data. However, as the bias is made lower to accurately model classification tasks, so is the accuracy of their parameters’ estimates, as each parameter is estimated from ever decreasing quantities of data. In this paper, we introduce the use of HDPs for accurate BNC parameter estimation even with lower bias. We conduct an extensive set of experiments on 68 standard datasets and demonstrate that our resulting classifiers perform very competitively with random forest in terms of prediction, while keeping the out-of-core capability and superior classification time.
引用
收藏
页码:1303 / 1331
页数:28
相关论文
共 50 条
  • [41] Learning Bayesian network classifiers using ant colony optimization
    Khalid M. Salama
    Alex A. Freitas
    Swarm Intelligence, 2013, 7 : 229 - 254
  • [42] Ensembles of Bayesian Network Classifiers Using Glaucoma Data and Expertise
    Ceccon, Stefano
    Garway-Heath, David
    Crabb, David
    Tucker, Allan
    ENSEMBLES IN MACHINE LEARNING APPLICATIONS, 2011, 373 : 133 - +
  • [43] Predicting Facial Biotypes Using Continuous Bayesian Network Classifiers
    Ruz, Gonzalo A.
    Araya-Diaz, Pamela
    COMPLEXITY, 2018,
  • [44] Novel and anomalous behavior detection using Bayesian network classifiers
    Benferhat, Salem
    Tabia, Karim
    SECRYPT 2008: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON SECURITY AND CRYPTOGRAPHY, 2008, : 13 - 20
  • [45] Learning Bayesian network classifiers using ant colony optimization
    Salama, Khalid M.
    Freitas, Alex A.
    SWARM INTELLIGENCE, 2013, 7 (2-3) : 229 - 254
  • [46] Learning Continuous Time Bayesian Network Classifiers Using MapReduce
    Villa, Simone
    Rossetti, Marco
    JOURNAL OF STATISTICAL SOFTWARE, 2014, 62 (03): : 1 - 25
  • [47] Efficient parameter learning for Bayesian Network classifiers following the Apache Spark Dataframes paradigm
    Akarepis, Ioannis
    Bompotas, Agorakis
    Makris, Christos
    KNOWLEDGE AND INFORMATION SYSTEMS, 2024, 66 (08) : 4437 - 4461
  • [48] Time series analysis using hierarchical neural network classifiers
    Dolenko, SA
    Orlov, YV
    Persiantsev, IG
    Shugai, JS
    PROCEEDINGS OF THE FIFTH JOINT CONFERENCE ON INFORMATION SCIENCES, VOLS 1 AND 2, 2000, : 908 - 911
  • [49] Inversion of hierarchical Bayesian models using Gaussian processes
    Lomakina, Ekaterina I.
    Paliwal, Saee
    Diaconescu, Andreea O.
    Brodersen, Kay H.
    Aponte, Eduardo A.
    Buhmann, Joachim M.
    Stephan, Klaas E.
    NEUROIMAGE, 2015, 118 : 133 - 145
  • [50] Update rules for parameter estimation in continuous time Bayesian network
    Shi, Dongyu
    You, Jinyuan
    PRICAI 2006: TRENDS IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2006, 4099 : 140 - 149