Considering diversity and accuracy simultaneously for ensemble pruning

被引:65
作者
Dai, Qun [1 ]
Ye, Rui [1 ]
Liu, Zhuan [1 ]
机构
[1] Nanjing Univ Aeronaut & Astronaut, Coll Comp Sci & Technol, Nanjing 211106, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Ensemble selection; Greedy ensemble pruning (GEP) algorithm; Simultaneous diversity & accuracy (SDAcc); Diversity-focused-two (DFTwo); Accuracy-reinforcement (AccRein); SELECTION; CLASSIFIERS; ALGORITHM; MACHINE; ERROR;
D O I
10.1016/j.asoc.2017.04.058
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Diversity among individual classifiers is widely recognized to be a key factor to successful ensemble selection, while the ultimate goal of ensemble pruning is to improve its predictive accuracy. Diversity and accuracy are two important properties of an ensemble. Existing ensemble pruning methods always consider diversity and accuracy separately. However, in contrast, the two closely interrelate with each other, and should be considered simultaneously. Accordingly, three new measures, i.e., Simultaneous Diversity & Accuracy, Diversity-Focused-Two and Accuracy-Reinforcement, are developed for pruning the ensemble by greedy algorithm. The motivation for Simultaneous Diversity & Accuracy is to consider the difference between the subensemble and the candidate classifier, and simultaneously, to consider the accuracy of both of them. With Simultaneous Diversity & Accuracy, those difficult samples are not given up so as to further improve the generalization performance of the ensemble. The inspiration of devising Diversity-Focused-Two stems from the cognition that ensemble diversity attaches more importance to the difference among the classifiers in an ensemble. Finally, the proposal of Accuracy-Reinforcement reinforces the concern about ensemble accuracy. Extensive experiments verified the effectiveness and efficiency of the proposed three pruning measures. Through the investigation of this work, it is found that by considering diversity and accuracy simultaneously for ensemble pruning, well-performed selective ensemble with superior generalization capability can be acquired, which is the scientific value of this paper. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:75 / 91
页数:17
相关论文
共 51 条
[41]   A multi-objective evolutionary algorithm-based ensemble optimizer for feature selection and classification with neural network models [J].
Tan, Choo Jun ;
Lim, Chee Peng ;
Cheah, Yu-N .
NEUROCOMPUTING, 2014, 125 :217-228
[42]   Selective fusion of heterogeneous classifiers [J].
Tsoumakas, Grigorios ;
Angelis, Lefteris ;
Vlahavas, Ioannis .
INTELLIGENT DATA ANALYSIS, 2005, 9 (06) :511-525
[43]   An Ensemble Pruning Primer [J].
Tsoumakas, Grigorios ;
Partalas, Ioannis ;
Vlahavas, Ioannis .
APPLICATIONS OF SUPERVISED AND UNSUPERVISED ENSEMBLE METHODS, 2009, 245 :1-13
[44]   Multistrategy ensemble learning: Reducing error by combining ensemble learning techniques [J].
Webb, GI ;
Zheng, ZJ .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2004, 16 (08) :980-991
[45]  
Witten I. H., 2005, DATA MINING PRACTICA
[46]   Somatic mutation detection using ensemble of flexible neural tree model [J].
Yang, Bin ;
Chen, Yuehui .
NEUROCOMPUTING, 2016, 179 :161-168
[47]   Self-organizing multiobjective optimization based on decomposition with neighborhood ensemble [J].
Zhang, Hu ;
Zhang, Xiujie ;
Gao, Xiao-Zhi ;
Song, Shenmin .
NEUROCOMPUTING, 2016, 173 :1868-1884
[48]  
Zhang Y, 2006, J MACH LEARN RES, V7, P1315
[49]  
Zhou DY, 2004, ADV NEUR IN, V16, P321
[50]   Ensembling neural networks: Many could be better than all [J].
Zhou, ZH ;
Wu, JX ;
Tang, W .
ARTIFICIAL INTELLIGENCE, 2002, 137 (1-2) :239-263