Distributed support vector machines

被引:60
作者
Navia-Vazquez, A. [1 ]
Gutierrez-Gonzalez, D. [1 ]
Parrado-Hernandez, E. [1 ]
Navarro-Abellan, J. J. [1 ]
机构
[1] Univ Carlos III Madrid, Dept Signal Theory & Commun, Madrid 28911, Spain
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2006年 / 17卷 / 04期
关键词
collaborative; compact; coopetitive; data mining; data privacy; distributed; information sharing; support vector machine;
D O I
10.1109/TNN.2006.875968
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A truly distributed (as opposed to parallelized) support vector machine (SVM) algorithm is presented. M-aining data are assumed to come from the same distribution and are locally stored in a number of different locations with processing capabilities (nodes). In several examples, it has been found that a reasonably small amount of information is interchanged among nodes to obtain an SVM solution, which is better than that obtained when classifiers are trained only with the local data and comparable (although a little bit worse) to that of the centralized approach (obtained when all the training data are available at the same place). We propose and analyze two distributed schemes: a "naive" distributed chunking approach, where raw data (support vectors) are communicated, and the more elaborated distributed semiparametric SVM, which aims at further reducing the total amount of information passed between nodes while providing a privacy-preserving mechanism for information sharing. We show the feasibility of our proposal by evaluating the performance of the algorithms in benchmarks with both synthetic and real-world datasets.
引用
收藏
页码:1091 / 1097
页数:7
相关论文
共 15 条
[1]  
[Anonymous], 2002, Least Squares Support Vector Machines
[2]  
[Anonymous], 1998, Encyclopedia of Biostatistics
[3]   TRAINING WITH NOISE IS EQUIVALENT TO TIKHONOV REGULARIZATION [J].
BISHOP, CM .
NEURAL COMPUTATION, 1995, 7 (01) :108-116
[4]  
Burges CJC, 1997, ADV NEUR IN, V9, P375
[5]  
Joachims T, 1999, ADVANCES IN KERNEL METHODS, P169
[6]   Weighted least squares training of support vector classifiers leading to compact and adaptive schemes [J].
Navia-Vázquez, A ;
Pérez-Cruz, F ;
Artés-Rodríguez, A ;
Figueiras-Vidal, AR .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2001, 12 (05) :1047-1059
[7]  
Osuna EE, 1999, ADVANCES IN KERNEL METHODS, P271
[8]  
PARK BH, 2003, DATA MINING HDB
[9]   Growing support vector classifiers with controlled complexity [J].
Parrado-Hernández, E ;
Mora-Jiménez, I ;
Arenas-García, J ;
Figueiras-Vidal, AR ;
Navia-Vázquez, A .
PATTERN RECOGNITION, 2003, 36 (07) :1479-1488
[10]   Convergence of the IRWLS procedure to the support vector machine solution [J].
Pérez-Cruz, F ;
Bousoño-Calzón, C ;
Artés-Rodríguez, A .
NEURAL COMPUTATION, 2005, 17 (01) :7-18