Parallel and Distributed Structured SVM Training

被引:7
作者
Jiang, Jiantong [1 ,2 ]
Wen, Zeyi [3 ]
Wang, Zeke [4 ]
He, Bingsheng [5 ]
Chen, Jian [6 ]
机构
[1] Univ Western Australia, Crawley, WA 6009, Australia
[2] Zhejiang Univ, Zhenjiang, Jiangsu 310027, Peoples R China
[3] Univ Western Australia, Crawley, WA 6009, Australia
[4] Zhejiang Univ, Hangzhou 310027, Zhejiang, Peoples R China
[5] Natl Univ Singapore, Singapore 119077, Singapore
[6] South China Univ Technol, Guangzhou 510006, Guangdong, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Training; Optimization; Support vector machines; Task analysis; Convergence; Synchronization; Natural languages; Parallel and distributed training; structured machine learning; support vector machines;
D O I
10.1109/TPDS.2021.3101155
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Structured Support Vector Machines (structured SVMs) are a fundamental machine learning algorithm, and have solid theoretical foundation and high effectiveness in applications such as natural language parsing and computer vision. However, training structured SVMs is very time-consuming, due to the large number of constraints and inferior convergence rates, especially for large training data sets. The high cost of training structured SVMs has hindered its adoption to new applications. In this article, we aim to improve the efficiency of structured SVMs by proposing a parallel and distributed solution (namely FastSSVM) for training structured SVMs building on top of MPI and OpenMP. FastSSVM exploits a series of optimizations (e.g., optimizations on data storage and synchronization) to efficiently use the resources of the nodes in a cluster and the cores of the nodes. Moreover, FastSSVM tackles the large constraint set problem by batch processing and addresses the slow convergence challenge by adapting stop conditions based on the improvement of each iteration. We theoretically prove that our solution is guaranteed to converge to a global optimum. A comprehensive experimental study shows that FastSSVM can achieve at least four times speedup over the existing solutions, and in some cases can achieve two to three orders of magnitude speedup.
引用
收藏
页码:1084 / 1096
页数:13
相关论文
共 49 条
  • [1] [Anonymous], 2010, PROC AISTATS
  • [2] [Anonymous], 2015, ARXIV150602620
  • [3] [Anonymous], 2013, HDB MATH
  • [4] [Anonymous], 2011, Acm T. Intel. Syst. Tec., DOI DOI 10.1145/1961189.1961199
  • [5] [Anonymous], 2004, Proceedings of the 21st International Conference on Machine Learning, DOI DOI 10.1145/1015330.1015341
  • [6] Belanger D, 2017, PR MACH LEARN RES, V70
  • [7] Boyd S., 2004, CONVEX OPTIMIZATION
  • [8] Efficient Large-Scale Structured Learning
    Branson, Steve
    Beijbom, Oscar
    Belongie, Serge
    [J]. 2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 1806 - 1813
  • [9] A MapReduce-based distributed SVM algorithm for binary classification
    Catak, Ferhat Ozgur
    Balaban, Mehmet Erdal
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2016, 24 (03) : 863 - 873
  • [10] Catanzaro Bryan, 2008, P 25 INT C MACH LEAR, P104, DOI DOI 10.1145/1390156.1390170