Parallel Dual Coordinate Descent Method for Large-scale Linear Classification in Multi-core Environments

被引:29
作者
Chiang, Wei-Lin [1 ]
Lee, Mu-Chu [1 ]
Lin, Chih-Jen [1 ]
机构
[1] Natl Taiwan Univ, Dept Comp Sci, Taipei, Taiwan
来源
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING | 2016年
关键词
dual coordinate descent; linear classification; multi-core computing;
D O I
10.1145/2939672.2939826
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Dual coordinate descent method is one of the most effective approaches for large-scale linear classification. However, its sequential design makes the parallelization difficult. In this work, we target at the parallelization in a multi-core environment. After pointing out difficulties faced in some existing approaches, we propose a new framework to parallelize the dual coordinate descent method. The key idea is to make the majority of all operations (gradient calculation here) parallelizable. The proposed framework is shown to be theoretically sound. Further, we demonstrate through experiments that the new framework is robust and efficient in a multi-core environment.
引用
收藏
页码:1485 / 1494
页数:10
相关论文
共 15 条
  • [1] [Anonymous], 1990, SUPPORT VECTOR LEARN
  • [2] [Anonymous], NIPS
  • [3] OpenMP: An industry standard API for shared-memory programming
    Dagum, L
    Menon, R
    [J]. IEEE COMPUTATIONAL SCIENCE & ENGINEERING, 1998, 5 (01): : 46 - 55
  • [4] Fan RE, 2005, J MACH LEARN RES, V6, P1889
  • [5] Fan RE, 2008, J MACH LEARN RES, V9, P1871
  • [6] Hildreth C., 1957, NAV RES LOG, V4, P79
  • [7] Hsieh C., 2008, ICML
  • [8] Hsieh C., 2015, ICML
  • [9] A simple decomposition method for support vector machines
    Hsu, CW
    Lin, CJ
    [J]. MACHINE LEARNING, 2002, 46 (1-3) : 291 - 314
  • [10] Joachims T., 1998, MAKING LARGE SCALE S