Incomplete data classification with voting based extreme learning machine

被引:18
作者
Yan, Yuan-Ting [1 ,2 ]
Zhang, Yan-Ping [1 ,2 ]
Chen, Jie [1 ,2 ]
Zhang, Yi-Wen [1 ,2 ]
机构
[1] Anhui Univ, Key Lab Intelligent Comp & Signal Proc, Minist Educ, Hefei 230039, Anhui, Peoples R China
[2] Anhui Univ, Sch Comp Sci & Technol, Hefei 230601, Anhui, Peoples R China
基金
中国国家自然科学基金;
关键词
Incomplete data; Extreme learning machine; Mutual information; Data subset; Weighted majority voting; NETWORKS; ENSEMBLE;
D O I
10.1016/j.neucom.2016.01.068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Extreme learning machine (ELM) was proposed as a new efficient learning algorithm for single-hidden layer feedforward neural networks (SLFN) in recent years. It is featured by its much faster training speed and better generalization performance over traditional SLFN learning techniques. However, ELM cannot deal directly with incomplete data which widely exists in real-world applications. In this paper, we propose a new algorithm to handle incomplete data with voting based extreme learning machine (V-ELMI). V-ELMI did not rely on any assumptions about missing values. It first obtains a group of data subsets according to the missing values of the training set. Then, it applies mutual information to measure the importance degree of each data subsets. After that, it trains a group of subclassifiers on these data subsets by applying ELM as base learning algorithm. Finally, for a given test sample with missing values, V-ELMI selects the subclassifiers whose input did not require the missing values to predict it. And final prediction is determined by weighted majority voting according to the mean value of the norms of the output weights and the importance degree of each available subclassifier. Experimental results on 15 UCI incomplete datasets and 5 UCI complete datasets have shown that, V-ELMI generally has better performance than the algorithms compared. Moreover, compared with the classification algorithms based on neural network ensemble (NNE), V-ELMI can greatly improve algorithm computational efficiency. (C) 2016 Elsevier B.V. All rights reserved.
引用
收藏
页码:167 / 175
页数:9
相关论文
共 33 条
[1]  
[Anonymous], 2002, Matrices: Theory and Applications
[2]  
[Anonymous], 2001, MISSING DATA
[3]  
[Anonymous], 2001, NEURAL NETWORKS COMP
[4]  
Benoit F., 2014, COMPUT STAT DATA AN, V71, P832
[5]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[6]   Voting based extreme learning machine [J].
Cao, Jiuwen ;
Lin, Zhiping ;
Huang, Guang-Bin ;
Liu, Nan .
INFORMATION SCIENCES, 2012, 185 (01) :66-77
[7]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[8]   A novel framework for imputation of missing values in databases [J].
Farhangfar, Alireza ;
Kurgan, Lukasz A. ;
Pedrycz, Witold .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 2007, 37 (05) :692-709
[9]   Research on collaborative negotiation for e-commerce. [J].
Feng, YQ ;
Lei, Y ;
Li, Y ;
Cao, RZ .
2003 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-5, PROCEEDINGS, 2003, :2085-2088
[10]   STOCHASTIC RELAXATION, GIBBS DISTRIBUTIONS, AND THE BAYESIAN RESTORATION OF IMAGES [J].
GEMAN, S ;
GEMAN, D .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1984, 6 (06) :721-741