Structural learning of mixed noisy-OR Bayesian networks 

被引:6
作者
Vomlel, Jiri [1 ]
Kratochvil, Vaclav [1 ]
Kratochvil, Frantisek [2 ]
机构
[1] Czech Acad Sci, Inst Informat Theory & Automat, Pod Vodarenskou Vezi 4, Prague 8, Czech Republic
[2] Palacky Univ Olomouc, Dept Asian Studies, Trida Svobody 26, Olomouc 77900, Czech Republic
关键词
Bayesian networks; Learning Bayesian networks; Noisy -OR model; Applications of Bayesian networks; Linguistics; Loanwords; MODEL;
D O I
10.1016/j.ijar.2023.108990
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper we discuss learning Bayesian networks whose conditional probability tables are either Noisy-OR models or general conditional probability tables. We refer to these models as Mixed Noisy-OR Bayesian Networks. To learn their structure, we modify the Bayesian Information Criterion used for standard Bayesian networks to reflect the number of parameters of a Noisy-OR model. We prove that the log-likelihood function of a Noisy-OR model has a unique maximum and adapt the EM-learning method for the leaky Noisy-OR model. We propose a structure learning algorithm that learns optimal Mixed Noisy-OR Bayesian Networks. We evaluate the proposed approach on synthetic data, where it performs substantially better than standard Bayesian networks. We perform experiments with Bipartite Noisy-OR Bayesian networks of different complexity to find out when the results of Mixed Noisy-OR Bayesian Networks are significantly better than the results of standard Bayesian networks and when they perform similarly. We also study how different penalties based on the number of model parameters affect the quality of the results. Finally, we apply the suggested approach to a problem from the domain of linguistics. Specifically, we use Mixed Noisy-OR Bayesian Networks to model the spread of loanwords in the South-East Asian Archipelago. We perform numerical experiments in which we compare the prediction ability of standard Bayesian networks with Mixed Noisy-OR Bayesian networks and test different pruning methods to reduce the number of parent sets considered.& COPY; 2023 Elsevier Inc. All rights reserved.
引用
收藏
页数:18
相关论文
共 29 条
[1]  
Almond R. G., 2015, Bayesian Networks in Educational Assessment, DOI DOI 10.1007/978-1-4939-2125-6
[2]  
Blust Robert, 2023, Zenodo, DOI 10.5281/ZENODO.7741197
[3]  
Chickering D. M., 2003, Journal of Machine Learning Research, V3, P507, DOI 10.1162/153244303321897717
[4]  
Chickering DM, 2004, J MACH LEARN RES, V5, P1287
[5]  
Cussens J., 2018, GOBNILP, Version 1.6.3
[6]   Bayesian Network Structure Learning with Integer Programming: Polytopes, Facets and Complexity [J].
Cussens, James ;
Jarvisalo, Matti ;
Korhonen, Janne H. ;
Bartlett, Mark .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2017, 58 :185-229
[7]   Entropy-based pruning for learning Bayesian networks using BIC [J].
de Campos, Cassio P. ;
Scanagatta, Mauro ;
Corani, Giorgio ;
Zaffalon, Marco .
ARTIFICIAL INTELLIGENCE, 2018, 260 :42-50
[8]  
Diez F.J., 2006, CISIAD0601 UNED
[9]   Efficient computation for the noisy MAX [J].
Diez, FJ ;
Galán, SF .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2003, 18 (02) :165-177
[10]  
Friedman N, 1996, UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, P252