Learning ELM-Tree from big data based on uncertainty reduction

被引:42
作者
Wang, Ran [1 ]
He, Yu-Lin [2 ]
Chow, Chi-Yin [1 ]
Ou, Fang-Fang [2 ]
Zhang, Jian [2 ]
机构
[1] City Univ Hong Kong, Dept Comp Sci, Kowloon, Hong Kong, Peoples R China
[2] Hebei Univ, Coll Math & Comp Sci, Key Lab Machine Learning & Computat Intelligence, Baoding 071002, Hebei, Peoples R China
基金
中国国家自然科学基金;
关键词
Big data classification; Decision tree; ELM-Tree; Extreme learning machine; Uncertainty reduction; PARALLEL; MACHINE; CLASSIFIERS; ATTRIBUTES; REGRESSION; INDUCTION;
D O I
10.1016/j.fss.2014.04.028
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
A challenge in big data classification is the design of highly parallelized learning algorithms. One solution to this problem is applying parallel computation to different components of a learning model. In this paper, we first propose an extreme learning machine tree (ELM-Tree) model based on the heuristics of uncertainty reduction. In the ELM-Tree model, information entropy and ambiguity are used as the uncertainty measures for splitting decision tree (DT) nodes. Besides, in order to resolve the over-partitioning problem in the DT induction, ELMs are embedded as the leaf nodes when the gain ratios of all the available splits are smaller than a given threshold. Then, we apply parallel computation to five components of the ELM-Tree model, which effectively reduces the computational time for big data classification. Experimental studies demonstrate the effectiveness of the proposed method. (C) 2014 Elsevier B.V. All rights reserved.
引用
收藏
页码:79 / 100
页数:22
相关论文
共 35 条
[1]  
[Anonymous], 2005, DATA MINING
[2]   Parameter selection algorithm with self adaptive growing neural network classifier for diagnosis issues [J].
Barakat, M. ;
Lefebvre, D. ;
Khalil, M. ;
Druaux, F. ;
Mustapha, O. .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2013, 4 (03) :217-233
[3]  
Ben-Haim Y, 2010, J MACH LEARN RES, V11, P849
[4]   Structural vibration suppression by using neural classifier with genetic algorithm [J].
Chen, Chuen-Jyh .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2012, 3 (03) :215-221
[5]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[6]  
Dong XL, 2013, PROC INT CONF DATA, P1245, DOI 10.1109/ICDE.2013.6544914
[7]   Smooth function approximation using neural networks [J].
Ferrari, S ;
Stengel, RF .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2005, 16 (01) :24-38
[8]   Technical note: Using model trees for classification [J].
Frank, E ;
Wang, Y ;
Inglis, S ;
Holmes, G ;
Witten, IH .
MACHINE LEARNING, 1998, 32 (01) :63-76
[9]   Additive logistic regression: A statistical view of boosting - Rejoinder [J].
Friedman, J ;
Hastie, T ;
Tibshirani, R .
ANNALS OF STATISTICS, 2000, 28 (02) :400-407
[10]   Functional trees [J].
Gama, J .
MACHINE LEARNING, 2004, 55 (03) :219-250