Design and Implementation of a Communication-Optimal Classifier for Distributed Kernel Support Vector Machines

被引:9
作者
You, Yang [1 ]
Demmel, James [1 ]
Czechowski, Kent [2 ]
Song, Le [2 ]
Vuduc, Rich [2 ]
机构
[1] Univ Calif Berkeley, Div Comp Sci, Berkeley, CA 94720 USA
[2] Georgia Tech, Coll Comp, Atlanta, GA 30332 USA
基金
美国国家科学基金会;
关键词
Distributed memory algorithms; communication-avoidance; statistical machine learning;
D O I
10.1109/TPDS.2016.2608823
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We consider the problem of how to design and implement communication-efficient versions of parallel kernel support vector machines, a widely used classifier in statistical machine learning, for distributed memory clusters and supercomputers. The main computational bottleneck is the training phase, in which a statistical model is built from an input data set. Prior to our study, the parallel isoefficiency of a state-of-the-art implementation scaled as W = Omega(P-3), where W is the problem size and P the number of processors; this scaling is worse than even a one-dimensional block row dense matrix vector multiplication, which has W = Omega(P-2). This study considers a series of algorithmic refinements, leading ultimately to a Communication-Avoiding SVM method that improves the isoefficiency to nearly W = Omega(P). We evaluate these methods on 96 to 1,536 processors, and show average speedups of 3 - 16 x ( 7 x on average) over Dis-SMO, and a 95 percent weak-scaling efficiency on six real-world datasets, with only modest losses in overall classification accuracy. The source code can be downloaded at [1].
引用
收藏
页码:974 / 988
页数:15
相关论文
共 37 条
  • [1] [Anonymous], 2008, P 25 INT C MACHINE L, DOI DOI 10.1145/1390156.1390170
  • [2] [Anonymous], 2003, Introduction to Parallel Computing
  • [3] Bertin-Mahieux T., 2011, P 12 INT SOC MUS INF, V2, P10, DOI DOI 10.7916/D8NZ8J07
  • [4] Parallel sequential minimal optimization for the training of support vector machines
    Cao, L. J.
    Keerthi, S. S.
    ong, Ch-Jin Ong
    Zhang, J. Q.
    Periyathamby, Uvaraj
    Fu, Xiu Ju
    Lee, H. P.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2006, 17 (04): : 1039 - 1049
  • [5] LIBSVM: A Library for Support Vector Machines
    Chang, Chih-Chung
    Lin, Chih-Jen
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
  • [6] Chang EdwardY., 2011, FDN OF LARGE SCALE M, P213, DOI DOI 10.1007/978-3-642-20429-6_10
  • [7] CORTES C, 1995, MACH LEARN, V20, P273, DOI 10.1023/A:1022627411411
  • [8] Perfect Strong Scaling Using No Additional Energy
    Demmel, James
    Gearhart, Andrew
    Lipshitz, Benjamin
    Schwartz, Oded
    [J]. IEEE 27TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2013), 2013, : 649 - 660
  • [9] Dongarra J., 2014, OP500 JUNE2014 LIST
  • [10] Fan RE, 2005, J MACH LEARN RES, V6, P1889