Classifier selection for majority voting

被引:403
作者
Ruta, Dymitr [1 ]
Gabrys, Bogdan [2 ]
机构
[1] Computational Intelligence Group, BT Exact Technologies, Orion Building, Martlesham Heath, Ipswich IP5 3 RE, pp12, Adastral Park
[2] Compl. Intelligence Research Group, Bournemouth University, Talbot Campus
关键词
Classifier fusion; Classifier selection; Diversity; Generalisation; Majority voting; Search algorithms;
D O I
10.1016/j.inffus.2004.04.008
中图分类号
学科分类号
摘要
Individual classification models are recently challenged by combined pattern recognition systems, which often show better performance. In such systems the optimal set of classifiers is first selected and then combined by a specific fusion method. For a small number of classifiers optimal ensembles can be found exhaustively, but the burden of exponential complexity of such search limits its practical applicability for larger systems. As a result, simpler search algorithms and/or selection criteria are needed to reduce the complexity. This work provides a revision of the classifier selection methodology and evaluates the practical applicability of diversity measures in the context of combining classifiers by majority voting. A number of search algorithms are proposed and adjusted to work properly with a number of selection criteria including majority voting error and various diversity measures. Extensive experiments carried out with 15 classifiers on 27 datasets indicate inappropriateness of diversity measures used as selection criteria in favour of the direct combiner error based search. Furthermore, the results prompted a novel design of multiple classifier systems in which selection and fusion are recurrently applied to a population of best combinations of classifiers rather than the individual best. The improvement of the generalisation performance of such system is demonstrated experimentally. © 2004 Elsevier B.V. All rights reserved.
引用
收藏
页码:63 / 81
页数:18
相关论文
共 40 条
[1]  
Baluja S., Population-based incremental learning: A method for integrating genetic, search based function optimization and competitive learning, Technical Report No. 163, 163, (1994)
[2]  
Clio S.-B., Pattern recognition with neural networks combined by genetic algorithms, Fuzzy Seta and Systems, 103, pp. 339-347, (1999)
[3]  
Davis L., Handbook of Genetic Algorithms, (1991)
[4]  
Drucker H., Cortes C., Jackel L.D., LeCun Y., Vapnik V., Boosting and other ensemble methods, Neural Computation, 6, pp. 1289-1301, (1994)
[5]  
Duda R.O., Hart P.E., Stork D.G., Pattern Classification, (2001)
[6]  
Pleiss J.L., Statistical Methods for Rates and Proportions, (1981)
[7]  
Giacinto G., Roli P., Methods for dynamic classifier selection, Proceedings of the 10th International Conference on Image - Analysis and Processing, pp. 659-664, (1999)
[8]  
Giacinto G., Roli F., A theoretical framework for dynamic classifier selection, Proceedings of the 15th International Conference on Pattern Recognition, 2, pp. 8-11, (2000)
[9]  
Giacinto G., Roli P., Design of effective neural network ensembles for image classification purposes, Image Vision and Computing Journal, 19, 9-10, pp. 669-707, (2001)
[10]  
Giacinto G., Roli F., Dynamic classifier selection based on multiple classifier behaviour, Pattern Recognition, 34, pp. 1879-1881, (2001)