Generalized core vector machines

被引:141
作者
Tsang, Ivor Wai-Hung
Kwok, James Tin-Yau
Zurada, Jacek M.
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci, Hong Kong, Hong Kong, Peoples R China
[2] Univ Louisville, Dept Elect & Comp Engn, Louisville, KY 40292 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2006年 / 17卷 / 05期
关键词
approximation algorithms; core vector machines (CVMs); kernel methods; minimum enclosing ball (MEB); quadratic programming; support vector machines (SVMs);
D O I
10.1109/TNN.2006.878123
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Kernel methods, such as the support vector machine (SVM), are often formulated as quadratic programming (QP) problems. However, given m training patterns, a naive implementation of the QP solver takes 0(m(3)) training time and at least 0(m(2)) space. Hence, scaling up these QPs is a major stumbling block in applying kernel methods on very large data sets, and a replacement of the naive method for finding the QP solutions is highly desirable. Recently, by using approximation algorithms for the minimum enclosing ball (MEB) problem, we proposed the core vector machine (CVM) algorithm that is much faster and can handle much larger data sets than existing SVM implementations,. However, the CVM can only be used with certain kernel functions and kernel methods. For example, the very popular support vector regression (SVR) cannot be used with the CVM. In this paper, we introduce the center-constrained MEB problem and subsequently extend the CVM algorithm. The generalized CVM algorithm can now be used with any linear/nonlinear kernel and can also be applied to kernel methods such as SVR and the ranking SVM. Moreover, like the original CVM, its asymptotic time complexity is again linear in m and its space complexity is independent of m. Experiments show that the generalized CVM has comparable performance with state-of-the-art SVM and SVR implementations, but is faster and produces fewer support vectors on very large data sets.
引用
收藏
页码:1126 / 1140
页数:15
相关论文
共 38 条
[1]  
Achlioptas D, 2002, ADV NEUR IN, V14, P335
[2]  
AGARWAL P, 2005, CURRENT TRENDS COMBI
[3]  
[Anonymous], 1979, Computers and Intractablity: A Guide to the Theoryof NP-Completeness
[4]  
[Anonymous], 1998, P 15 INT C MACHINE L
[5]  
[Anonymous], 1998, Encyclopedia of Biostatistics
[6]  
BADOIU M, 2002, OPTIMAL CORE SETS BA
[7]   The use of the area under the roc curve in the evaluation of machine learning algorithms [J].
Bradley, AP .
PATTERN RECOGNITION, 1997, 30 (07) :1145-1159
[8]  
CHANG CC, 2004, LIBSVM LIB SUPPORT V
[9]   An improved conjugate gradient scheme to the solution of least squares SVM [J].
Chu, W ;
Ong, CJ ;
Keerthi, SS .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2005, 16 (02) :498-501
[10]   A parallel mixture of SVMs for very large scale problems [J].
Collobert, R ;
Bengio, S ;
Bengio, Y .
NEURAL COMPUTATION, 2002, 14 (05) :1105-1114