A Systematic Comparison of Supervised Classifiers

被引:145
作者
Amancio, Diego Raphael [1 ]
Comin, Cesar Henrique [2 ]
Casanova, Dalcimar [2 ]
Travieso, Gonzalo [2 ]
Bruno, Odemir Martinez [2 ]
Rodrigues, Francisco Aparecido [1 ]
Costa, Luciano da Fontoura [2 ]
机构
[1] Univ Sao Paulo, Inst Math & Comp Sci, Sao Paulo, Brazil
[2] Univ Sao Paulo, Sao Carlos Inst Phys, Sao Paulo, Brazil
来源
PLOS ONE | 2014年 / 9卷 / 04期
基金
巴西圣保罗研究基金会;
关键词
PATTERN-RECOGNITION; PARAMETER OPTIMIZATION; VECTOR MACHINES; NEURAL-NETWORKS; ALGORITHMS; MATRICES;
D O I
10.1371/journal.pone.0094137
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Pattern recognition has been employed in a myriad of industrial, commercial and academic applications. Many techniques have been devised to tackle such a diversity of applications. Despite the long tradition of pattern recognition research, there is no technique that yields the best classification in all scenarios. Therefore, as many techniques as possible should be considered in high accuracy applications. Typical related works either focus on the performance of a given algorithm or compare various classification methods. In many occasions, however, researchers who are not experts in the field of machine learning have to deal with practical classification tasks without an in-depth knowledge about the underlying parameters. Actually, the adequate choice of classifiers and parameters in such practical circumstances constitutes a long-standing problem and is one of the subjects of the current paper. We carried out a performance study of nine well-known classifiers implemented in the Weka framework and compared the influence of the parameter configurations on the accuracy. The default configuration of parameters in Weka was found to provide near optimal performance for most cases, not including methods such as the support vector machine (SVM). In addition, the k-nearest neighbor method frequently allowed the best accuracy. In certain conditions, it was possible to improve the quality of SVM by more than 20% with respect to their default parameter configuration.
引用
收藏
页数:14
相关论文
共 57 条
  • [31] Automated, High Accuracy Classification of Parkinsonian Disorders: A Pattern Recognition Approach
    Marquand, Andre F.
    Filippone, Maurizio
    Ashburner, John
    Girolami, Mark
    Mourao-Miranda, Janaina
    Barker, Gareth J.
    Williams, Steven C. R.
    Leigh, P. Nigel
    Blain, Camilla R. V.
    [J]. PLOS ONE, 2013, 8 (07):
  • [32] GENERATING CORRELATION-MATRICES
    MARSAGLIA, G
    OLKIN, I
    [J]. SIAM JOURNAL ON SCIENTIFIC AND STATISTICAL COMPUTING, 1984, 5 (02): : 470 - 475
  • [33] The support vector machine under test
    Meyer, D
    Leisch, F
    Hornik, K
    [J]. NEUROCOMPUTING, 2003, 55 (1-2) : 169 - 186
  • [34] Comparison of fast nearest neighbour classifiers for handwritten character recognition
    Mico, L
    Oncina, J
    [J]. PATTERN RECOGNITION LETTERS, 1998, 19 (3-4) : 351 - 356
  • [35] Machine learning of molecular electronic properties in chemical compound space
    Montavon, Gregoire
    Rupp, Matthias
    Gobre, Vivekanand
    Vazquez-Mayagoitia, Alvaro
    Hansen, Katja
    Tkatchenko, Alexandre
    Mueller, Klaus-Robert
    von Lilienfeld, O. Anatole
    [J]. NEW JOURNAL OF PHYSICS, 2013, 15
  • [36] Moschitti A, 2003, LECT NOTES COMPUT SC, V2633, P420
  • [37] Pérez O, 2007, LECT NOTES COMPUT SC, V4528, P192
  • [38] The Validation and Assessment of Machine Learning: A Game of Prediction from High-Dimensional Data
    Pers, Tune H.
    Albrechtsen, Anders
    Holst, Claus
    Sorensen, Thorkild I. A.
    Gerds, Thomas A.
    [J]. PLOS ONE, 2009, 4 (08):
  • [39] A quantitative study of experimental evaluations of neural network learning algorithms: Current research practice
    Prechelt, L
    [J]. NEURAL NETWORKS, 1996, 9 (03) : 457 - 462
  • [40] Meta-learning for evolutionary parameter optimization of classifiers
    Reif, Matthias
    Shafait, Faisal
    Dengel, Andreas
    [J]. MACHINE LEARNING, 2012, 87 (03) : 357 - 380