Cascade generalization

被引:183
作者
Gama, J [1 ]
Brazdil, P [1 ]
机构
[1] Univ Porto, FEP, LIACC, P-8234150 Porto, Portugal
关键词
multiple models; constructive induction; combining classifiers; merging classifiers;
D O I
10.1023/A:1007652114878
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Using multiple classifiers for increasing learning accuracy is an active research area. In this paper we present two related methods for merging classifiers. The first method, Cascade Generalization, couples classifiers loosely. It belongs to the family of stacking algorithms. The basic idea of Cascade Generalization is to use sequentially the set of classifiers, at each step performing an extension of the original data by the insertion of new attributes. The new attributes are derived from the probability class distribution given by a base classifier. This constructive step extends the representational language for the high level classifiers, relaxing their bias. The second method exploits tight coupling of classifiers, by applying Cascade Generalization locally. At each iteration of a divide and conquer algorithm, a reconstruction of the instance space occurs by the addition of new attributes. Each new attribute represents the probability that an example belongs to a class given by a base classifier. We have implemented three Local Generalization Algorithms. The first merges a linear discriminant with a decision tree, the second merges a naive Bayes with a decision tree, and the third merges a linear discriminant and a naive Bayes with a decision tree. All the algorithms show an increase of performance, when compared with the corresponding single models. Cascade also outperforms other methods for combining classifiers, like Stacked Generalization, and competes well against Boosting at statistically significant confidence levels.
引用
收藏
页码:315 / 343
页数:29
相关论文
共 34 条
[1]  
Ali KM, 1996, MACH LEARN, V24, P173, DOI 10.1007/BF00058611
[2]  
[Anonymous], MACHINE LEARNING
[3]  
[Anonymous], 1997, THESIS U MASSACHUSET
[4]  
[Anonymous], 1991, Advances in Neural Information Processing Systems
[5]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[6]  
Blake C., 1999, Uci repository of machine learning data sets
[7]   SmcHD1, containing a structural-maintenance-of-chromosomes hinge domain, has a critical role in X inactivation [J].
Blewitt, Marnie E. ;
Gendrel, Anne-Valerie ;
Pang, Zhenyi ;
Sparrow, Duncan B. ;
Whitelaw, Nadia ;
Craig, Jeffrey M. ;
Apedaile, Anwyn ;
Hilton, Douglas J. ;
Dunwoodie, Sally L. ;
Brockdorff, Neil ;
Kay, Graham F. ;
Whitelaw, Emma .
NATURE GENETICS, 2008, 40 (05) :663-669
[8]  
Breiman L, 1998, ANN STAT, V26, P801
[9]   RECURSIVE AUTOMATIC BIAS SELECTION FOR CLASSIFIER CONSTRUCTION [J].
BRODLEY, CE .
MACHINE LEARNING, 1995, 20 (1-2) :63-94
[10]  
BRODLEY CE, 1995, MACH LEARN, V19, P45, DOI 10.1007/BF00994660