On robust randomized neural networks for regression: a comprehensive review and evaluation

被引:10
作者
Freire, Ananda L. [1 ]
Rocha-Neto, Ajalmar R. [1 ]
Barreto, Guilherme A. [2 ]
机构
[1] Fed Inst Ceara, Dept Comp Sci, Fortaleza, Ceara, Brazil
[2] Fed Univ Ceara UFC, Dept Teleinformat Engn, Fortaleza, Ceara, Brazil
关键词
Randomized neural networks; Robustness; Outliers; Numerical stability; Regularization; EXTREME LEARNING-MACHINE; RIDGE-REGRESSION; CLASSIFICATION; CLASSIFIERS; ALGORITHMS; PLASTICITY; PREDICTION;
D O I
10.1007/s00521-020-04994-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Data from real-world regression problems are quite often contaminated with outliers. In order to efficiently handle such undesirable samples, robust parameter estimation methods have been incorporated into randomized neural network (RNN) models, usually replacing the ordinary least squares (OLS) method. Despite recent successful applications to outlier-contaminated scenarios, significant issues remain unaddressed in the design of reliable outlier-robust RNN models for regression tasks. For example, the number of hidden neurons impacts directly on the norm of the estimated output weights, since the OLS method will rely on an ill-conditioned hidden-layer output matrix. Another design concern involves the high sensitivity of RNNs to the randomization of the hidden layer weights, an issue that can be suitably handled, e.g., by intrinsic plasticity techniques. Bearing these concerns in mind, we describe several ideas introduced in previous works concerning the design of RNN models that are both robust to outliers and numerically stable. A comprehensive evaluation of their performances is carried out across several benchmarking regression datasets taking into account accuracy, weight norms, and training time as figures of merit.
引用
收藏
页码:16931 / 16950
页数:20
相关论文
共 62 条
[1]   The multivariate least-trimmed squares estimator [J].
Agullo, Jose ;
Croux, Christophe ;
Van Aelst, Stefan .
JOURNAL OF MULTIVARIATE ANALYSIS, 2008, 99 (03) :311-338
[2]   RELATIONSHIP BETWEEN VARIABLE SELECTION AND DATA AUGMENTATION AND A METHOD FOR PREDICTION [J].
ALLEN, DM .
TECHNOMETRICS, 1974, 16 (01) :125-127
[3]   1-Norm extreme learning machine for regression and multiclass classification using Newton method [J].
Balasundaram, S. ;
Gupta, Deepak ;
Kapil .
NEUROCOMPUTING, 2014, 128 :4-14
[4]   A Robust Extreme Learning Machine for pattern classification with outliers [J].
Barreto, Guilherme A. ;
Barros, Ana Luiza B. P. .
NEUROCOMPUTING, 2016, 176 :3-13
[5]  
Barreto GA, 2015, NEURAL PROCESS LETT, V42, P119, DOI 10.1007/s11063-014-9393-2
[6]  
Barros ALBP, 2013, LECT NOTES COMPUT SC, V8073, P588, DOI 10.1007/978-3-642-40846-5_59
[7]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[8]  
Beliakov G, 2011, ARXIV11100169 CORR
[9]   Distributed optimization and statistical learning via the alternating direction method of multipliers [J].
Boyd S. ;
Parikh N. ;
Chu E. ;
Peleato B. ;
Eckstein J. .
Foundations and Trends in Machine Learning, 2010, 3 (01) :1-122
[10]   Fast iteratively reweighted least squares algorithms for analysis-based sparse reconstruction [J].
Chen, Chen ;
He, Lei ;
Li, Hongsheng ;
Huang, Junzhou .
MEDICAL IMAGE ANALYSIS, 2018, 49 :141-152