Regression-Based Hyperparameter Learning for Support Vector Machines

被引:12
作者
Peng, Shili [1 ]
Wang, Wenwu [2 ]
Chen, Yinli [1 ]
Zhong, Xueling [1 ]
Hu, Qinghua [3 ]
机构
[1] Guangdong Univ Finance, Sch Internet Finance & Informat Engn, Guangzhou 510521, Peoples R China
[2] Univ Surrey, Ctr Vis Speech & Signal Proc, Guildford GU2 7XH, Surrey, England
[3] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300072, Peoples R China
基金
中国国家自然科学基金;
关键词
Support vector machines; Task analysis; Classification algorithms; Bayes methods; Uncertainty; Prediction algorithms; Training; Hyperparameter optimization; maximum margin classification; regression; support vector machine (SVM); OUT CROSS-VALIDATION; GAUSSIAN-PROCESSES; KERNEL; CLASSIFICATION; NOISE;
D O I
10.1109/TNNLS.2023.3321685
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unification of classification and regression is a major challenge in machine learning and has attracted increasing attentions from researchers. In this article, we present a new idea for this challenge, where we convert the classification problem into a regression problem, and then use the methods in regression to solve the problem in classification. To this end, we leverage the widely used maximum margin classification algorithm and its typical representative, support vector machine (SVM). More specifically, we convert SVM into a piecewise linear regression task and propose a regression-based SVM (RBSVM) hyperparameter learning algorithm, where regression methods are used to solve several key problems in classification, such as learning of hyperparameters, calculation of prediction probabilities, and measurement of model uncertainty. To analyze the uncertainty of the model, we propose a new concept of model entropy, where the leave-one-out prediction probability of each sample is converted into entropy, and then used to quantify the uncertainty of the model. The model entropy is different from the classification margin, in the sense that it considers the distribution of all samples, not just the support vectors. Therefore, it can assess the uncertainty of the model more accurately than the classification margin. In the case of the same classification margin, the farther the sample distribution is from the classification hyperplane, the lower the model entropy. Experiments show that our algorithm (RBSVM) provides higher prediction accuracy and lower model uncertainty, when compared with state-of-the-art algorithms, such as Bayesian hyperparameter search and gradient-based hyperparameter learning algorithms.
引用
收藏
页码:18799 / 18813
页数:15
相关论文
共 56 条
[1]   Well-M3N: A Maximum-Margin Approach to Unsupervised Structured Prediction [J].
Abidi, Shaukat ;
Piccardi, Massimo ;
Tsang, Ivor W. ;
Williams, Mary-Anne .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2019, 3 (06) :427-439
[2]   A Semiproximal Support Vector Machine Approach for Binary Multiple Instance Learning [J].
Avolio, Matteo ;
Fuduli, Antonio .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (08) :3566-3577
[3]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[4]  
Bergstra JamesDaniel Yamins David Cox., 2013, MAKING SCI MODEL SEA, P115, DOI [10.5555/3042817.3042832, DOI 10.5555/3042817.3042832]
[5]   Efficient approximate leave-one-out cross-validation for kernel logistic regression [J].
Cawley, Gavin C. ;
Talbot, Nicola L. C. .
MACHINE LEARNING, 2008, 71 (2-3) :243-264
[6]   Fast exact leave-one-out cross-validation of sparse least-squares support vector machines [J].
Cawley, GC ;
Talbot, NLC .
NEURAL NETWORKS, 2004, 17 (10) :1467-1475
[7]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[8]   Choosing multiple parameters for support vector machines [J].
Chapelle, O ;
Vapnik, V ;
Bousquet, O ;
Mukherjee, S .
MACHINE LEARNING, 2002, 46 (1-3) :131-159
[9]   Privacy-Preserving Multi-Class Support Vector Machine Model on Medical Diagnosis [J].
Chen, Yange ;
Mao, Qinyu ;
Wang, Baocang ;
Duan, Pu ;
Zhang, Benyu ;
Hong, Zhiyong .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (07) :3342-3353
[10]   Automatic Optimization of Hyperparameters Using Associative Self-adapting Structures [J].
Czaplak, Szymon ;
Horzyk, Adrian .
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,