Optimization-based Extreme Learning Machine with Multi-kernel Learning Approach for Classification

被引:11
作者
Cao, Le-le [1 ]
Huang, Wen-bing [1 ]
Sun, Fu-chun [1 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Tsinghua Natl Lab Informat Sci & Technol TNList, Beijing 100084, Peoples R China
来源
2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2014年
关键词
multi-kernel extreme learning machine (MK-ELM); extreme learning machine (ELM); multi-kernel learning (MKL); optimization-based ELM; SimpleMKL; KERNEL; REGRESSION; NETWORKS;
D O I
10.1109/ICPR.2014.613
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The optimization method based extreme learning machine (optimization-based ELM) is generalized from single-hidden-layer feed-forward neural networks (SLFNs) by making use of kernels instead of neuron-alike hidden nodes. This approach is known for its high scalability, low computational complexity, and mild optimization constrains. The multi-kernel learning (MKL) framework SimpleMKL iteratively determines the combination of kernels by gradient descent wrapping a standard support vector machine (SVM) solver. SimpleMKL can be applied to many kinds of supervised learning problems to receive a more stable performance with rapid convergence speed. This paper proposes a new approach: MK-ELM (multi-kernel extreme learning machine) that applies SimpleMKL framework to the optimization-based ELM algorithm. The performance analysis on binary classification problems with various scales shows that MK-ELM tends to achieve the best generalization performance as well as being the most insensitive to parameters comparing to optimization-based ELM and SimpleMKL. As a result, MK-ELM can be implemented in real applications easily.
引用
收藏
页码:3564 / 3569
页数:6
相关论文
共 34 条
[1]  
[Anonymous], 2010, P 18 EUR S ART NEUR
[2]  
[Anonymous], OPTIMISATION CONTINU
[3]  
[Anonymous], 1981, PRACTICAL METHODS OP
[4]  
[Anonymous], P 21 INT C MACH LEAR
[5]   The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network [J].
Bartlett, PL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1998, 44 (02) :525-536
[6]  
Blake C. L., 1998, Uci repository of machine learning databases
[7]   Optimization problems with perturbations: A guided tour [J].
Bonnans, JF ;
Shapiro, A .
SIAM REVIEW, 1998, 40 (02) :228-264
[8]  
Canu S., 2003, EA4108 LITIS INSA DE
[9]   Choosing multiple parameters for support vector machines [J].
Chapelle, O ;
Vapnik, V ;
Bousquet, O ;
Mukherjee, S .
MACHINE LEARNING, 2002, 46 (1-3) :131-159
[10]  
DeCoste D., 2000, Proceedings. KDD-2000. Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, P345, DOI 10.1145/347090.347165