An ensemble based on neural networks with random weights for online data stream regression

被引:13
作者
de Almeida, Ricardo [1 ,2 ]
Goh, Yee Mey [2 ]
Monfared, Radmehr [2 ]
Steiner, Maria Teresinha Arns [1 ]
West, Andrew [2 ]
机构
[1] Pontificia Univ Catolica Parana, Ind & Syst Engn Grad Program, R Imaculada Conceicao 1155, BR-80215901 Curitiba, Parana, Brazil
[2] Loughborough Univ, Wolfson Sch Mech Elect & Mfg Engn, Loughborough LE11 3TU, Leics, England
关键词
Ensembles; Data stream regression; Neural networks with random weights; Hyperparameter adjustment; CONCEPT DRIFT; EVOLVING DATA; ALGORITHM;
D O I
10.1007/s00500-019-04499-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most information sources in the current technological world are generating data sequentially and rapidly, in the form of data streams. The evolving nature of processes may often cause changes in data distribution, also known as concept drift, which is difficult to detect and causes loss of accuracy in supervised learning algorithms. As a consequence, online machine learning algorithms that are able to update actively according to possible changes in the data distribution are required. Although many strategies have been developed to tackle this problem, most of them are designed for classification problems. Therefore, in the domain of regression problems, there is a need for the development of accurate algorithms with dynamic updating mechanisms that can operate in a computational time compatible with today's demanding market. In this article, the authors propose a new bagging ensemble approach based on neural network with random weights for online data stream regression. The proposed method improves the data prediction accuracy as well as minimises the required computational time compared to a recent algorithm for online data stream regression from literature. The experiments are carried out using four synthetic datasets to evaluate the algorithm's response to concept drift, along with four benchmark datasets from different industries. The results indicate improvement in data prediction accuracy, effectiveness in handling concept drift, and much faster updating times compared to the existing available approach. Additionally, the use of design of experiments as an effective tool for hyperparameter tuning is demonstrated.
引用
收藏
页码:9835 / 9855
页数:21
相关论文
共 46 条
[1]   Fast decorrelated neural network ensembles with random weights [J].
Alhamdoosh, Monther ;
Wang, Dianhui .
INFORMATION SCIENCES, 2014, 264 :104-117
[2]  
Bifet A, 2009, KDD-09: 15TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, P139
[3]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[4]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[5]  
Brown G., 2005, Information Fusion, V6, P5, DOI 10.1016/j.inffus.2004.04.004
[6]   A review on neural networks with random weights [J].
Cao, Weipeng ;
Wang, Xizhao ;
Ming, Zhong ;
Gao, Jinzhu .
NEUROCOMPUTING, 2018, 275 :278-287
[7]   Genetic algorithms for hyperparameter optimization in predictive business process monitoring [J].
Di Francescomarino, Chiara ;
Dumas, Marlon ;
Federici, Marco ;
Ghidini, Chiara ;
Maria Maggi, Fabrizio ;
Rizzi, Williams ;
Simonetto, Luca .
INFORMATION SYSTEMS, 2018, 74 :67-83
[8]   An online learning neural network ensembles with random weights for regression of sequential data stream [J].
Ding, Jinliang ;
Wang, Haitao ;
Li, Chuanbao ;
Chai, Tianyou ;
Wang, Junwei .
SOFT COMPUTING, 2017, 21 (20) :5919-5937
[9]   Incremental Learning of Concept Drift in Nonstationary Environments [J].
Elwell, Ryan ;
Polikar, Robi .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2011, 22 (10) :1517-1531
[10]  
Fan W., 2004, SIGKDD Conference on Knowledge Discovery and Data Mining, P128, DOI [10.1145/1014052.1014069, DOI 10.1145/1014052.1014069]