Faster Support Vector Machines

被引:0
作者
Schlag S. [1 ]
Schmitt M. [1 ]
Schulz C. [2 ]
机构
[1] Karlsruhe Institute of Technology, Faculty of Computer Science
[2] Heidelberg University, Faculty of Mathematics and Computer Science
来源
ACM Journal of Experimental Algorithmics | 2021年 / 26卷
基金
欧洲研究理事会;
关键词
graph clustering; kernel; model parameter optimization; multilevel training; Support vector machines;
D O I
10.1145/3484730
中图分类号
学科分类号
摘要
The time complexity of support vector machines (SVMs) prohibits training on huge datasets with millions of data points. Recently, multilevel approaches to train SVMs have been developed to allow for time-efficient training on huge datasets. While regular SVMs perform the entire training in one - time-consuming - optimization step, multilevel SVMs first build a hierarchy of problems decreasing in size that resemble the original problem and then train an SVM model for each hierarchy level, benefiting from the solved models of previous levels. We present a faster multilevel support vector machine that uses a label propagation algorithm to construct the problem hierarchy. Extensive experiments indicate that our approach is up to orders of magnitude faster than the previous fastest algorithm while having comparable classification quality. For example, already one of our sequential solvers is on average a factor 15 faster than the parallel ThunderSVM algorithm, while having similar classification quality.. © 2021 ACM.
引用
收藏
相关论文
共 61 条
[1]  
Aizerman M.A., Braverman E.M., Rozoner L.I., Theoretical foundations of the potential function method in pattern recognition learning, Autom. Rem. Contr., 25, pp. 821-837, (1964)
[2]  
Allen Zhu Z., Chen W., Wang G., Zhu C., Chen Z., P-packSVM: Parallel primal grAdient desCent kernel SVM, Proceedings of the 9th IEEE International Conference on Data Mining, pp. 677-686, (2009)
[3]  
Alpaydin E., Introduction to Machine Learning (2nd ed.), (2010)
[4]  
Boser B.E., Guyon I.M., Vapnik V.N., A training algorithm for optimal margin classifiers, Proceedings of the 5th Annual Workshop on Computational Learning Theory (COLT'92). ACM, pp. 144-152, (1992)
[5]  
Buluc A., Meyerhenke H., Safro I., Sanders P., Schulz C., Recent advances in graph partitioning, Algorithm Engineering-Selected Results and Surveys. Springer, pp. 117-158, (2016)
[6]  
Chang C., Lin C., LIBSVM: A library for support vector machines, ACM Trans. Intell. Syst. Technol., 2, 3, pp. 271-2727, (2011)
[7]  
Chang C., Lin C., LIBSVM FAQ, (2019)
[8]  
Chang E.Y., Zhu K., Wang H., Bai H., Li J., Qiu Z., Cui H., Parallelizing support vectormachines on distributed computers, Proceedings of the 21st Annual Conference on Neural Information Processing Systems. Curran Associates, Inc., pp. 257-264, (2007)
[9]  
Chapelle O., Vapnik V., Bousquet O., Mukherjee S., Choosing multiple parameters for support vector machines, Mach. Learn., 46, 1-3, pp. 131-159, (2002)
[10]  
Chen C., Li X., Nasreddine Belkacem A., Qiao Z., Dong E., Tan W., Shin D., The mixed kernel function SVM-based point cloud classification, Int. J. Precis. Eng. Manuf., 20, 5, pp. 737-747, (2019)