Learning mixtures of polynomials of multidimensional probability densities from data using B-spline interpolation

被引:11
作者
Lopez-Cruz, Pedro L. [1 ]
Bielza, Concha [1 ]
Larranaga, Pedro [1 ]
机构
[1] Univ Politecn Madrid, Fac Informat, Computat Intelligence Grp, Dept Inteligencia Artificial, E-28660 Madrid, Spain
关键词
Mixtures of polynomials; Interpolation; Density estimation; Bayesian classifiers; BAYESIAN NETWORKS; NAIVE-BAYES; TRUNCATED EXPONENTIALS; BANDWIDTH SELECTION; KNOT SELECTION; REGRESSION; DISCRETIZATION; CLASSIFIERS; ASSUMPTION; EXTENSION;
D O I
10.1016/j.ijar.2013.09.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Non-parametric density estimation is an important technique in probabilistic modeling and reasoning with uncertainty. We present a method for learning mixtures of polynomials (Mops) approximations of one-dimensional and multidimensional probability densities from data. The method is based on basis spline interpolation, where a density is approximated as a linear combination of basis splines. We compute maximum likelihood estimators of the mixing coefficients of the linear combination. The Bayesian information criterion is used as the score function to select the order of the polynomials and the number of pieces of the Mop. The method is evaluated in two ways. First, we test the approximation fitting. We sample artificial datasets from known one-dimensional and multidimensional densities and learn MoP approximations from the datasets. The quality of the approximations is analyzed according to different criteria, and the new proposal is compared with MoPs learned with Lagrange interpolation and mixtures of truncated basis functions. Second, the proposed method is used as a non-parametric density estimation technique in Bayesian classifiers. Two of the most widely studied Bayesian classifiers, the naive Bayes and tree-augmented naive Bayes classifiers, are implemented and compared. Results on real datasets show that the non-parametric Bayesian classifiers using MOPS are comparable to the kernel density-based Bayesian classifiers. We provide a free R package implementing the proposed methods. (C) 2013 Elsevier Inc. All rights reserved.
引用
收藏
页码:989 / 1010
页数:22
相关论文
共 77 条
  • [1] Alcalá-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
  • [2] [Anonymous], LECT NOTES COMPUTER
  • [3] [Anonymous], INT J APPRO IN PRESS
  • [4] [Anonymous], 2012, P 6 EUR WORKSH PROB
  • [5] [Anonymous], 1978, A Practical Guide to Splines
  • [6] [Anonymous], KERNSMOOTH FUNCTIONS
  • [7] [Anonymous], P 2 S INT COMP SICO2
  • [8] [Anonymous], IMA J APPL MATH
  • [9] [Anonymous], KS KERNEL SMOOTHING
  • [10] [Anonymous], CUBATURE ADAPTIVE MU