A Maximally Split and Relaxed ADMM for Regularized Extreme Learning Machines

被引:32
作者
Lai, Xiaoping [1 ]
Cao, Jiuwen [1 ]
Huang, Xiaofeng [2 ]
Wang, Tianlei [1 ]
Lin, Zhiping [3 ]
机构
[1] Hangzhou Dianzi Univ, Inst Informat & Control, Hangzhou 310018, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Commun Engn, Hangzhou 310018, Peoples R China
[3] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
关键词
Convex functions; Convergence; Training; Acceleration; Standards; Approximation algorithms; Complexity theory; Alternating direction method of multipliers (ADMM); computational complexity; convergence rate; extreme learning machine (ELM); parallel algorithm; ALTERNATING DIRECTION METHOD; DIRECT EXTENSION; CONVERGENCE; OPTIMIZATION; MULTIPLIERS; REGRESSION; ALGORITHM;
D O I
10.1109/TNNLS.2019.2927385
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the salient features of the extreme learning machine (ELM) is its fast learning speed. However, in a big data environment, the ELM still suffers from an overly heavy computational load due to the high dimensionality and the large amount of data. Using the alternating direction method of multipliers (ADMM), a convex model fitting problem can be split into a set of concurrently executable subproblems, each with just a subset of model coefficients. By maximally splitting across the coefficients and incorporating a novel relaxation technique, a maximally split and relaxed ADMM (MS-RADMM), along with a scalarwise implementation, is developed for the regularized ELM (RELM). The convergence conditions and the convergence rate of the MS-RADMM are established, which exhibits linear convergence with a smaller convergence ratio than the unrelaxed maximally split ADMM. The optimal parameter values of the MS-RADMM are obtained and a fast parameter selection scheme is provided. Experiments on ten benchmark classification data sets are conducted, the results of which demonstrate the fast convergence and parallelism of the MS-RADMM. Complexity comparisons with the matrix-inversion-based method in terms of the numbers of multiplication and addition operations, the computation time and the number of memory cells are provided for performance evaluation of the MS-RADMM.
引用
收藏
页码:1899 / 1913
页数:15
相关论文
共 52 条
[1]  
[Anonymous], FOUND TRENDS MACH LE
[2]  
Antoniou A., 2007, PRACTICAL OPTIMIZATI
[3]  
Antsaklis P., 2007, LINEAR SYSTEMS PRIME
[4]   Generalized symmetric ADMM for separable convex optimization [J].
Bai, Jianchao ;
Li, Jicheng ;
Xu, Fengmin ;
Zhang, Hongchao .
COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2018, 70 (01) :129-170
[5]   The ''independent components'' of natural scenes are edge filters [J].
Bell, AJ ;
Sejnowski, TJ .
VISION RESEARCH, 1997, 37 (23) :3327-3338
[6]  
Bertsekas D., 2015, CONVEX OPTIMIZATION, DOI ISBN1-886529-28-0
[7]  
Burden RL., 2011, Numerical Analysis, V9th ed
[8]   On the convergence of the direct extension of ADMM for three-block separable convex minimization models with one strongly convex function [J].
Cai, Xingju ;
Han, Deren ;
Yuan, Xiaoming .
COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2017, 66 (01) :39-73
[9]   Extreme Learning Machine With Affine Transformation Inputs in an Activation Function [J].
Cao, Jiuwen ;
Zhang, Kai ;
Yong, Hongwei ;
Lai, Xiaoping ;
Chen, Badong ;
Lin, Zhiping .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (07) :2093-2107
[10]   Extreme learning machine and adaptive sparse representation for image classification [J].
Cao, Jiuwen ;
Zhang, Kai ;
Luo, Minxia ;
Yin, Chun ;
Lai, Xiaoping .
NEURAL NETWORKS, 2016, 81 :91-102