One-Class SVM Probabilistic Outputs

被引:10
作者
Que, Zhongyi [1 ,2 ]
Lin, Chih-Jen [1 ,3 ]
机构
[1] Natl Taiwan Univ, Dept Comp Sci & Informat Engn, Taipei 106, Taiwan
[2] Univ Tokyo, Dept Math Informat, Tokyo 1138654, Japan
[3] Mohamed Bin Zayed Univ Artificial Intelligence, Dept Machine Learning, Abu Dhabi, U Arab Emirates
关键词
Support vector machines; Probabilistic logic; Standards; Optimization; Training data; Training; Gaussian distribution; One-class support vector machine (SVM); outlier detection; Platt scaling; probability estimation; SUPPORT; SCORES;
D O I
10.1109/TNNLS.2024.3395148
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One-class support vector machine (SVM) is an extension of SVM to handle unlabeled data. As a mature technique for outlier detection, one-class SVM has been widely used in many applications. However, similar to standard two-class SVM, the design of one-class SVM does not give probabilistic outputs. For two-class SVM, some methods have been proposed to effectively obtain probabilistic outputs, but due to the difficulty of no-label information, less attention has been paid to one-class SVM. Our aim in this work is to propose some practically viable techniques to generate probabilistic outputs for one-class SVM. We investigate existing methods for two-class SVM and explain why they may not be suitable for one-class SVM. Due to the lack of label information, we think a feasible setting is to have probabilities mimic to the decision values of training data. Based on this principle, we propose several new methods. Detailed experiments on both artificial and real-world data demonstrate the effectiveness of the proposed methods.
引用
收藏
页码:6244 / 6256
页数:13
相关论文
共 25 条
[1]   AN EMPIRICAL DISTRIBUTION FUNCTION FOR SAMPLING WITH INCOMPLETE INFORMATION [J].
AYER, M ;
BRUNK, HD ;
EWING, GM ;
REID, WT ;
SILVERMAN, E .
ANNALS OF MATHEMATICAL STATISTICS, 1955, 26 (04) :641-647
[2]   Estimating Outlier Score Probabilities [J].
Bauder, Richard A. ;
Khoshgoftaar, Taghi M. .
2017 IEEE 18TH INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION (IEEE IRI 2017), 2017, :559-568
[3]   A comparison of efficient approximations for a weighted sum of chi-squared random variables [J].
Bodenham, Dean A. ;
Adams, Niall M. .
STATISTICS AND COMPUTING, 2016, 26 (04) :917-928
[4]  
Bounsiar A., 2014, 2014 International Conference on Information Science Applications (ICISA), P1, DOI DOI 10.1109/ICISA.2014.6847419
[5]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[6]   Probabilistic Novelty Detection With Support Vector Machines [J].
Clifton, Lei ;
Clifton, David A. ;
Zhang, Yang ;
Watkinson, Peter ;
Tarassenko, Lionel ;
Yin, Hujun .
IEEE TRANSACTIONS ON RELIABILITY, 2014, 63 (02) :455-467
[7]   Import Vector Domain Description: A Kernel Logistic One-Class Learning Algorithm [J].
Decherchi, Sergio ;
Rocchia, Walter .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (07) :1722-1729
[8]   Converting SVDD scores into probability estimates: Application to outlier detection [J].
El Azami, Meriem ;
Lartizien, Carole ;
Canu, Stephane .
NEUROCOMPUTING, 2017, 268 :64-75
[9]  
Gao J, 2006, IEEE DATA MINING, P212
[10]  
Glazer A., 2012, P ADV NEUR INF PROC, V25