Latent classification models for binary data

被引:7
作者
Langseth, Helge [1 ]
Nielsen, Thomas D. [2 ]
机构
[1] Norwegian Univ Sci & Technol, Dept Informat & Comp Sci, N-7491 Trondheim, Norway
[2] Aalborg Univ, Dept Comp Sci, DK-9220 Aalborg, Denmark
关键词
Classification; Binary images; Bayesian networks; Variational inference; NAIVE BAYES;
D O I
10.1016/j.patcog.2009.05.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the simplest, and yet most consistently well-performing set of classifiers is the naive Bayes models (a special class of Bayesian network models). However, these models rely on the (naive) assumption that all the attributes used to describe an instance are conditionally independent given the class of that instance. To relax this independence assumption, we have in previous work proposed a family of models, called latent classification models (LCMs). LCMs are defined for continuous domains and generalize the naive Bayes model by using latent variables to model class-conditional dependencies between the attributes. In addition to providing good classification accuracy, the LCM has several appealing properties, including a relatively small parameter space making it less susceptible to over-fitting. In this paper we take a first step towards generalizing LCMs to hybrid domains, by proposing an LCM for domains with binary attributes. We present algorithms for learning the proposed model, and we describe a variational approximation-based inference procedure. Finally, we empirically compare the accuracy of the proposed model to the accuracy of other classifiers for a number of different domains, including the problem of recognizing symbols in black and white images. (C) 2009 Elsevier Ltd. All rights reserved.
引用
收藏
页码:2724 / 2736
页数:13
相关论文
共 43 条
[1]  
AHA DW, 1991, MACH LEARN, V6, P37, DOI 10.1007/BF00153759
[2]  
[Anonymous], 2004, WILEY SER PROB STAT
[3]  
[Anonymous], 1987, Latent variable models and factors analysis
[4]  
[Anonymous], 1997, MACHINE LEARNING, MCGRAW-HILL SCIENCE/ENGINEERING/MATH
[5]  
BISHOP CM, 2002, P 7 VAL INT M BAYES, P98
[6]  
Bouchard G., 2007, NIPS 2007 WORKSHOP A
[7]  
Bressan M, 2002, LECT NOTES ARTIF INT, V2527, P1
[8]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[9]   A BAYESIAN METHOD FOR THE INDUCTION OF PROBABILISTIC NETWORKS FROM DATA [J].
COOPER, GF ;
HERSKOVITS, E .
MACHINE LEARNING, 1992, 9 (04) :309-347
[10]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38