A comparative study of classifier ensembles for bankruptcy prediction

被引:126
作者
Tsai, Chih-Fong [1 ]
Hsu, Yu-Feng [2 ]
Yen, David C. [3 ]
机构
[1] Natl Cent Univ, Dept Informat Management, Jhongli, Taiwan
[2] Natl Sun Yat Sen Univ, Dept Informat Management, Kaohsiung 80424, Taiwan
[3] SUNY Coll Oneonta, Sch Business & Econ, Oneonta, NY 13820 USA
关键词
Bankruptcy prediction; Credit scoring; Classifier ensembles; Data mining; Machine learning; NEURAL-NETWORKS; DECISION TREES;
D O I
10.1016/j.asoc.2014.08.047
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The aim of bankruptcy prediction in the areas of data mining and machine learning is to develop an effective model which can provide the higher prediction accuracy. In the prior literature, various classification techniques have been developed and studied, in/with which classifier ensembles by combining multiple classifiers approach have shown their outperformance over many single classifiers. However, in terms of constructing classifier ensembles, there are three critical issues which can affect their performance. The first one is the classification technique actually used/adopted, and the other two are the combination method to combine multiple classifiers and the number of classifiers to be combined, respectively. Since there are limited, relevant studies examining these aforementioned disuses, this paper conducts a comprehensive study of comparing classifier ensembles by three widely used classification techniques including multilayer perceptron (MLP) neural networks, support vector machines (SVM), and decision trees (DT) based on two well-known combination methods including bagging and boosting and different numbers of combined classifiers. Our experimental results by three public datasets show that DT ensembles composed of 80-100 classifiers using the boosting method perform best. The Wilcoxon signed ranked test also demonstrates that DT ensembles by boosting perform significantly different from the other classifier ensembles. Moreover, a further study over a real-world case by a Taiwan bankruptcy dataset was conducted, which also demonstrates the superiority of DT ensembles by boosting over the others. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:977 / 984
页数:8
相关论文
共 35 条
[1]   Bankruptcy forecasting:: An empirical comparison of AdaBoost and neural networks [J].
Alfaro, Esteban ;
Garcia, Noelia ;
Gamez, Matias ;
Elizondo, David .
DECISION SUPPORT SYSTEMS, 2008, 45 (01) :110-122
[2]  
[Anonymous], 2006, International Journal of Hybrid Intelligent Systems, DOI [10.3233/HIS-2006-3104, DOI 10.3233/HIS-2006-3104]
[3]   SmcHD1, containing a structural-maintenance-of-chromosomes hinge domain, has a critical role in X inactivation [J].
Blewitt, Marnie E. ;
Gendrel, Anne-Valerie ;
Pang, Zhenyi ;
Sparrow, Duncan B. ;
Whitelaw, Nadia ;
Craig, Jeffrey M. ;
Apedaile, Anwyn ;
Hilton, Douglas J. ;
Dunwoodie, Sally L. ;
Brockdorff, Neil ;
Kay, Graham F. ;
Whitelaw, Emma .
NATURE GENETICS, 2008, 40 (05) :663-669
[4]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[5]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[6]   An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization [J].
Dietterich, TG .
MACHINE LEARNING, 2000, 40 (02) :139-157
[7]  
Freund Y., 1996, INT C MACH LEARN ICM, V6, P148, DOI DOI 10.5555/3091696.3091715
[8]   A theoretical and experimental analysis of linear combiners for multiple classifier systems [J].
Fumera, G ;
Roli, F .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2005, 27 (06) :942-956
[9]   DESIGN AND EVOLUTION OF MODULAR NEURAL-NETWORK ARCHITECTURES [J].
HAPPEL, BLM ;
MURRE, JMJ .
NEURAL NETWORKS, 1994, 7 (6-7) :985-1004
[10]  
Haykin S., 1999, Neural Networks: A Comprehensive Foundation, DOI DOI 10.1017/S0269888998214044