Optimization extreme learning machine with ν regularization

被引:11
作者
Ding Xiao-jian [4 ]
Lan Yuan [2 ]
Zhang Zhi-feng [3 ]
Xu Xin [1 ]
机构
[1] Sci & Technol Informat Syst Engn Lab, Nanjing 210007, Jiangsu, Peoples R China
[2] Taiyuan Univ Technol, Sch Mech Engn, Minist Educ Adv Transducers & Intelligent Control, Key Lab, Taiyuan 030024, Shanxi, Peoples R China
[3] Zhengzhou Univ Light Ind, Software Coll, Zhengzhou 450002, Henan, Peoples R China
[4] Huawei Technol Co Ltd, Nanjing 211100, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
nu-optimization extreme learning machine; Classification; Parameter selection; CLASSIFICATION; NETWORKS; REGRESSION; NEURONS;
D O I
10.1016/j.neucom.2016.05.114
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The problem of choosing error penalty parameter C for optimization extreme learning machine (OELM) is that it can take any positive value for different applications and it is therefore hard to choose correctly. In this paper, we reformulated OELM to take a new regularization parameter nu (nu-OELM) which is inspired by Scholkopf et al. The regularization in terms of nu is bounded between 0 and 1, and is easier to interpret as compared to C. This paper shows that: (1) nu-OELM and nu-SVM have similar dual optimization formulation, but nu-OELM has less optimization constraints due to its special capability of class separation and (2) experiment results on both artificial and real binary classification problems show that nu-OELM tends to achieve better generalization performance than nu-SVM, OELM and other popular machine learning approaches, and it is computationally efficient on high dimension data sets. Additionally, the optimal parameter nu in nu-OELM can be easily selected from few candidates. (C) 2017 Elsevier B.V. All rights reserved.
引用
收藏
页码:11 / 19
页数:9
相关论文
共 28 条
[1]   Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays [J].
Alon, U ;
Barkai, N ;
Notterman, DA ;
Gish, K ;
Ybarra, S ;
Mack, D ;
Levine, AJ .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 1999, 96 (12) :6745-6750
[2]  
[Anonymous], P 18 EUR S ART NEUR
[3]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[4]   Silicon spiking neurons for hardware implementation of extreme learning machines [J].
Basu, Arindam ;
Shuo, Sun ;
Zhou, Hongming ;
Lim, Meng Hiot ;
Huang, Guang-Bin .
NEUROCOMPUTING, 2013, 102 :125-134
[5]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[6]   An ELM-based model for affective analogical reasoning [J].
Cambria, Erik ;
Gastaldo, Paolo ;
Bisio, Federica ;
Zunino, Rodolfo .
NEUROCOMPUTING, 2015, 149 :443-455
[7]  
Canu S., 2005, Perception Systmes et Information
[8]   Training ν-support vector classifiers:: Theory and algorithms [J].
Chang, CC ;
Lin, CJ .
NEURAL COMPUTATION, 2001, 13 (09) :2119-2147
[9]  
CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
[10]   Approximate statistical tests for comparing supervised classification learning algorithms [J].
Dietterich, TG .
NEURAL COMPUTATION, 1998, 10 (07) :1895-1923