Parsimonious regularized extreme learning machine based on orthogonal transformation

被引:13
作者
Zhao, Yong-Ping [1 ]
Wang, Kang-Kang [1 ]
Li, Ye-Bo [2 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Mech Engn, Nanjing 210094, Jiangsu, Peoples R China
[2] AVIC Aeroengine Control Res Inst, Wuxi 214063, Peoples R China
基金
中国国家自然科学基金;
关键词
Extreme learning machine; Sparseness; Tikhonov regularization; Orthogonal transformation; Condition number; FEEDFORWARD NETWORKS; NEURAL-NETWORKS; CLASSIFICATION; OPTIMIZATION; REGRESSION; IDENTIFICATION; APPROXIMATION; ALGORITHM; SELECTION; ELM;
D O I
10.1016/j.neucom.2014.12.046
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, two parsimonious algorithms were proposed to sparsify extreme learning machine (ELM), i.e., constructive parsimonious ELM (CP-ELM) and destructive parsimonious ELM (DP-ELM). In this paper, the ideas behind CP-ELM and DP-ELM are extended to the regularized ELM (RELM), thus obtaining CP-RELM and DP-RELM. For CP-RELM(DP-RELM), there are two schemes to realize it, viz. CP-RELM-I and CP-RELM-II(DP-RELM-I and DP-RELM-II). Generally speaking, CP-RELM-II(DP-RELM-II) outperforms CP-RELM-I(DP-RELM-I) in terms of parsimoniousness. Under nearly the same generalization, compared with CP-ELM (DP-ELM), CP-RELM-II(DP-RELM-II) usually needs fewer hidden nodes. In addition, different from CP-ELM and DP-ELM, for CP-RELM and DP-RELM the number of candidate hidden nodes may be larger than the number of training samples, which assists the selection of much better hidden nodes for constructing more compact networks. Finally, eleven benchmark data sets divided into two groups are utilized to do experiments and the usefulness of the proposed algorithms is reported. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:280 / 296
页数:17
相关论文
共 44 条
[1]  
[Anonymous], 2001, Pattern Classification
[2]  
[Anonymous], 2017, Matrix Analysis and Applications
[3]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[4]   AN ALGORITHM FOR RLS IDENTIFICATION OF PARAMETERS THAT VARY QUICKLY WITH TIME [J].
BOBROW, JE ;
MURRAY, W .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1993, 38 (02) :351-354
[5]  
Bontempi G., 1998, P EUR C MACH LEARN, P292
[6]  
CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
[7]  
Deng WY, 2010, NEURAL NETW WORLD, V20, P317
[8]   Regularized Extreme Learning Machine [J].
Deng, Wanyu ;
Zheng, Qinghua ;
Chen, Lin .
2009 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DATA MINING, 2009, :389-395
[9]   Least angle regression - Rejoinder [J].
Efron, B ;
Hastie, T ;
Johnstone, I ;
Tibshirani, R .
ANNALS OF STATISTICS, 2004, 32 (02) :494-499
[10]   Evolutionary selection extreme learning machine optimization for regression [J].
Feng, Guorui ;
Qian, Zhenxing ;
Zhang, Xinpeng .
SOFT COMPUTING, 2012, 16 (09) :1485-1491