Sample selection-based hierarchical extreme learning machine

被引:4
作者
Xu, Xinzheng [1 ,2 ,3 ]
Li, Shan [1 ,2 ]
Liang, Tianming [1 ]
Sun, Tongfeng [1 ,3 ]
机构
[1] China Univ Min & Technol, Sch Comp Sci & Technol, Xuzhou 221116, Jiangsu, Peoples R China
[2] Minist Educ, Engn Res Ctr Min Digital, Xuzhou 221116, Jiangsu, Peoples R China
[3] Lanzhou Jiaotong Univ, Key Lab Optotechnol & Intelligent Control, Minist Educ, Lanzhou 730070, Gansu, Peoples R China
关键词
Sample selection; Fuzzy C-means clustering; Condensed nearest neighbour; Hierarchical extreme learning machine; ELM; ALGORITHM; DEEP;
D O I
10.1016/j.neucom.2019.10.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large amounts of training data in machine learning can keep the accuracy high to a certain extent, but the time costs are high because of the exorbitant amount of data and their dimensionality. Therefore, how to simultaneously select the most useful training data set and extract the main features of the samples, especially for image data, are essential problems that urgently need to be solved in the field of large-scale machine learning. Herein, a training sample selection method that is based on the fuzzy c-means clustering algorithm (FCM) is proposed for the problems. It first utilises condensed nearest neighbour (CNN) to make a preliminary selection of training samples. Then, it utilises the FCM to get the centres of the selected data, and, finally, it effectively condenses the sample using a compression parameter. Meanwhile, considering the critical influence of the sample features on the classification model, this paper selects the hierarchical extreme learning machine (H-ELM) model to better solve the classification task. Based on this, the paper presents the FCM-CNN-H-ELM framework for data classification, which combines FCM-Based CNN and H-ELM. The results of the experiments show that the proposed training sample selection method and classification framework can guarantee consistent, even higher, prediction results with a small number of training samples, and significantly reduce the training time. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:95 / 102
页数:8
相关论文
共 45 条
[11]   MR-ELM: a MapReduce-based framework for large-scale ELM training in big data era [J].
Chen, Jiaoyan ;
Chen, Huajun ;
Wan, Xiangyi ;
Zheng, Guozhou .
NEURAL COMPUTING & APPLICATIONS, 2016, 27 (01) :101-110
[12]   Learning Rotation-Invariant Convolutional Neural Networks for Object Detection in VHR Optical Remote Sensing Images [J].
Cheng, Gong ;
Zhou, Peicheng ;
Han, Junwei .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2016, 54 (12) :7405-7415
[13]   Empirical kernel map-based multilayer extreme learning machines for representation learning [J].
Chi-Man Vong ;
Chen, Chuangquan ;
Wong, Pak-Kin .
NEUROCOMPUTING, 2018, 310 :265-276
[14]   Sampling for Big Data: A Tutorial [J].
Cormode, Graham ;
Duffield, Nick .
PROCEEDINGS OF THE 20TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'14), 2014, :1975-1975
[15]   An incremental prototype set building technique [J].
Devi, VS ;
Murty, MN .
PATTERN RECOGNITION, 2002, 35 (02) :505-513
[16]   Extreme learning machine and its applications [J].
Ding, Shifei ;
Xu, Xinzheng ;
Nie, Ru .
NEURAL COMPUTING & APPLICATIONS, 2014, 25 (3-4) :549-556
[17]   Dissimilarity-Based Sparse Subset Selection [J].
Elhamifar, Ehsan ;
Sapiro, Guillermo ;
Sastry, S. Shankar .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (11) :2182-2197
[18]   Democratic instance selection: A linear complexity instance selection algorithm based on classifier ensemble concepts [J].
Garcia-Osorio, Cesar ;
de Haro-Garcia, Aida ;
Garcia-Pedrajas, Nicolas .
ARTIFICIAL INTELLIGENCE, 2010, 174 (5-6) :410-441
[19]   Kernel ELM and CNN based Facial Age Estimation [J].
Gurpinar, Furkan ;
Kaya, Heysem ;
Dibeklioglu, Hamdi ;
Salah, Albert Ali .
PROCEEDINGS OF 29TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, (CVPRW 2016), 2016, :785-791
[20]  
Hao Hong-Wei, 2007, Acta Automatica Sinica, V33, P1247, DOI 10.1360/aas-007-1247