Stochastic configuration network ensembles with selective base models

被引:30
作者
Huang, Changqin [1 ,2 ]
Li, Ming [3 ]
Wang, Dianhui [4 ,5 ]
机构
[1] Zhejiang Normal Univ, Key Lab Intelligent Educ Technol & Applicat Zheji, Jinhua, Zhejiang, Peoples R China
[2] South China Normal Univ, Sch Informat Technol Educ, Guangzhou, Peoples R China
[3] Zhejiang Normal Univ, Dept Educ Technol, Jinhua, Zhejiang, Peoples R China
[4] Northeastern Univ, State Key Lab Synthet Automat Proc Ind, Shenyang, Liaoning, Peoples R China
[5] La Trobe Univ, Dept Comp Sci & Informat Technol, Melbourne, Vic, Australia
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Stochastic configuration networks; Randomized learner models; Neural network ensemble; Educational data analytics; NEURAL-NETWORKS; APPROXIMATION; ALGORITHMS;
D O I
10.1016/j.neunet.2021.01.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Studies have demonstrated that stochastic configuration networks (SCNs) have good potential for rapid data modeling because of their sufficient adequate learning power, which is theoretically guaranteed. Empirical studies have verified that the learner models produced by SCNs can usually achieve favorable test performance in practice but more in-depth theoretical analysis of their generalization power would be useful for constructing SCN-based ensemble models with enhanced generalization capacities. In particular, given a collection of independently developed SCN-based learner models, it is useful to select certain base learners that can potentially obtain preferable test results rather than considering all of the base models together, before simply taking their average in order to build an effective ensemble model. In this study, we propose a novel framework for building SCN ensembles by exploring key factors that might potentially affect the generalization performance of the base model. Under a mild assumption, we provide a comprehensive theoretical framework for examining a learner model's generalization error, as well as formulating a novel indicator that contains measurement information for the training errors, output weights, and a hidden layer output matrix, which can be used by our proposed algorithm to find a subset of appropriate base models from a pool of randomized learner models. A toy example of one-dimensional function approximation, a case study for developing a predictive model for forecasting student learning performance, and two large-scale data sets were used in our experiments. The experimental results indicate that our proposed method has some remarkable advantages for building ensemble models. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页码:106 / 118
页数:13
相关论文
共 36 条
[1]   Fast decorrelated neural network ensembles with random weights [J].
Alhamdoosh, Monther ;
Wang, Dianhui .
INFORMATION SCIENCES, 2014, 264 :104-117
[2]   An ensemble of dynamic neural network identifiers for fault detection and isolation of gas turbine engines [J].
Amozegar, M. ;
Khorasani, K. .
NEURAL NETWORKS, 2016, 76 :106-121
[3]   Clustering ensembles of neural network models [J].
Bakker, B ;
Heskes, T .
NEURAL NETWORKS, 2003, 16 (02) :261-269
[4]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[5]  
Brown G, 2005, J MACH LEARN RES, V6, P1621
[6]   Regularized Negative Correlation Learning for Neural Network Ensembles [J].
Chen, Huanhuan ;
Yao, Xin .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (12) :1962-1979
[7]   Ensemble Neural Networks (ENN): A gradient-free stochastic method [J].
Chen, Yuntian ;
Chang, Haibin ;
Meng, Jin ;
Zhang, Dongxiao .
NEURAL NETWORKS, 2019, 110 :170-185
[8]  
Dattorro J., 2010, Convex optimization & Euclidean distance geometry
[9]   Ensemble methods in machine learning [J].
Dietterich, TG .
MULTIPLE CLASSIFIER SYSTEMS, 2000, 1857 :1-15
[10]   BOOSTING A WEAK LEARNING ALGORITHM BY MAJORITY [J].
FREUND, Y .
INFORMATION AND COMPUTATION, 1995, 121 (02) :256-285