Direct Kernel Perceptron (DKP): Ultra-fast kernel ELM-based classification with non-iterative closed-form weight calculation

被引:38
作者
Fernandez-Delgado, Manuel [1 ]
Cernadas, Eva [1 ]
Barro, Senen [1 ]
Ribeiro, Jorge [3 ]
Neves, Jose [2 ]
机构
[1] Univ Santiago de Compostela, Ctr Invest Tecnoloxias Informac USC CITIUS, La Coruna 15782, Spain
[2] Univ Minho, DI CCTC Dept Informat, P-4710057 Braga, Portugal
[3] Viana do Castelo Polytech Inst, Sch Technol & Management, Viana Do Castelo, Portugal
关键词
Kernel-based classification; Extreme learning machine; Support vector machine; Analytical weight calculation; Closed-form solution; Margin maximization; Parallel Delta rule; EXTREME LEARNING-MACHINE; NEURAL-NETWORK;
D O I
10.1016/j.neunet.2013.11.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Direct Kernel Perceptron (DKP) (Fernandez-Delgado et al., 2010) is a very simple and fast kernel-based classifier, related to the Support Vector Machine (SVM) and to the Extreme Learning Machine (ELM) (Huang, Wang, & Lan, 2011), whose a-coefficients are calculated directly, without any iterative training, using an analytical closed-form expression which involves only the training patterns. The DKP, which is inspired by the Direct Parallel Perceptron, (Auer et al., 2008), uses a Gaussian kernel and a linear classifier (perceptron). The weight vector of this classifier in the feature space minimizes an error measure which combines the training error and the hyperplane margin, without any tunable regularization parameter. This weight vector can be translated, using a variable change, to the a-coefficients, and both are determined without iterative calculations. We calculate solutions using several error functions, achieving the best trade-off between accuracy and efficiency with the linear function. These solutions for the a coefficients can be considered alternatives to the ELM with a new physical meaning in terms of error and margin: in fact, the linear and quadratic DKP are special cases of the two-class ELM when the regularization parameter C takes the values C = 0 and C = infinity. The linear DKP is extremely efficient and much faster (over a vast collection of 42 benchmark and real-life data sets) than 12 very popular and accurate classifiers including SVM, Multi-Layer Perceptron, Adaboost, Random Forest and Bagging of RPART decision trees, Linear Discriminant Analysis, K-Nearest Neighbors, ELM, Probabilistic Neural Networks, Radial Basis Function neural networks and Generalized ART. Besides, despite its simplicity and extreme efficiency, DKP achieves higher accuracies than 7 out of 12 classifiers, exhibiting small differences with respect to the best ones (SVM, ELM, Adaboost and Random Forest), which are much slower. Thus, the DKP provides an easy and fast way to achieve classification accuracies which are not too far from the best one for a given problem. The C and Matlab code of DKP are freely available.(1) (C) 2013 Elsevier-Ltd. All rights reserved.
引用
收藏
页码:60 / 71
页数:12
相关论文
共 39 条
  • [1] [Anonymous], 2013, R PROJECT STAT COMPU
  • [2] [Anonymous], EXTREME LEARNING MAC
  • [3] [Anonymous], 2007, Uci machine learning repository
  • [4] [Anonymous], 2013, KERNEL FISHER DISCRI
  • [5] A learning rule for very simple universal approximators consisting of a single layer of perceptrons
    Auer, Peter
    Burgsteiner, Harald
    Maass, Wolfgang
    [J]. NEURAL NETWORKS, 2008, 21 (05) : 786 - 795
  • [6] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [7] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [8] Classification of honeybee pollen using a multiscale texture filtering scheme
    Carrión, P
    Cernadas, E
    Gálvez, J
    Damián, M
    de Sá-Otero, P
    [J]. MACHINE VISION AND APPLICATIONS, 2004, 15 (04) : 186 - 193
  • [9] Chang C. C., 2008, LIBSVM LIB SUPPORT V
  • [10] SUPPORT-VECTOR NETWORKS
    CORTES, C
    VAPNIK, V
    [J]. MACHINE LEARNING, 1995, 20 (03) : 273 - 297