On the relationship between deterministic and probabilistic directed Graphical models: From Bayesian networks to recursive neural networks

被引:12
作者
Baldi, P [1 ]
Rosen-Zvi, M
机构
[1] Univ Calif Irvine, Sch Informat & Comp Sci, Irvine, CA 92697 USA
[2] Univ Calif Irvine, Inst Genom & Bioinformat, Irvine, CA 92697 USA
[3] Hebrew Univ Jerusalem, Sch Comp Sci & Engn, IL-91904 Jerusalem, Israel
基金
美国国家科学基金会; 美国国家卫生研究院;
关键词
Bayesian networks; belief propagation; recursive neural networks; recurrent neural networks; constraint networks graphical models;
D O I
10.1016/j.neunet.2005.07.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning methods that can handle variable-size structured data such as sequences and graphs include Bayesian networks (BNs) and Recursive Neural Networks (RNNs). In both classes of models, the data is modeled using a set of observed and hidden variables associated with the nodes of a directed acyclic graph. In BNs, the conditional relationships between parent and child variables are probabilistic, whereas in RNNs they are deterministic and parameterized by neural networks. Here, we study the formal relationship between both classes of models and show that when the source nodes variables are observed, RNNs can be viewed as limits, both in distribution and probability, of BNs with local conditional distributions that have vanishing covariance matrices and converge to delta functions. Conditions for uniform convergence are also given together with an analysis of the behavior and exactness of Belief Propagation (BP) in 'deterministic' BNs. Implications for the design of mixed architectures and the corresponding inference algorithms are briefly discussed. (c) 2005 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1080 / 1086
页数:7
相关论文
共 16 条
[1]  
[Anonymous], 1998, TUTORIAL LEARNING BA
[2]   The principled design of large-scale recursive neural network architectures-DAG-RNNs and the protein structure prediction problem [J].
Baldi, P ;
Pollastri, G .
JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 4 (04) :575-602
[3]   Hybrid modeling, HMM/NN architectures, and protein applications [J].
Baldi, P ;
Chauvin, Y .
NEURAL COMPUTATION, 1996, 8 (07) :1541-1565
[4]  
BALDI P, 2005, RELATIONSHIP DETERMI
[5]  
BARBER D, 2000, ADV NEURAL INFORMATI, V12
[6]  
Billingsley P., 1986, PROBABILITY MEASURE
[7]  
BOZHENA R, 2001, EPSILON CUTSET EFFEC
[8]  
Dechter R., 2003, CONSTRAINT PROCESSIN
[9]   A general framework for adaptive processing of data structures [J].
Frasconi, P ;
Gori, M ;
Sperduti, A .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (05) :768-786
[10]  
Goller C, 1996, IEEE IJCNN, P347, DOI 10.1109/ICNN.1996.548916