Optimised one-class classification performance

被引:0
作者
Oliver Urs Lenz
Daniel Peralta
Chris Cornelis
机构
[1] Ghent University,Department of Applied Mathematics, Computer Science and Statistics
[2] Ghent University—imec,IDLab, Department of Information Technology
来源
Machine Learning | 2022年 / 111卷
关键词
Data descriptors; Hyperparameter optimisation; Novelty detection; One-class classification; Semi-supervised outlier detection;
D O I
暂无
中图分类号
学科分类号
摘要
We provide a thorough treatment of one-class classification with hyperparameter optimisation for five data descriptors: Support Vector Machine (SVM), Nearest Neighbour Distance (NND), Localised Nearest Neighbour Distance (LNND), Local Outlier Factor (LOF) and Average Localised Proximity (ALP). The hyperparameters of SVM and LOF have to be optimised through cross-validation, while NND, LNND and ALP allow an efficient form of leave-one-out validation and the reuse of a single nearest-neighbour query. We experimentally evaluate the effect of hyperparameter optimisation with 246 classification problems drawn from 50 datasets. From a selection of optimisation algorithms, the recent Malherbe–Powell proposal optimises the hyperparameters of all data descriptors most efficiently. We calculate the increase in test AUROC and the amount of overfitting as a function of the number of hyperparameter evaluations. After 50 evaluations, ALP and SVM significantly outperform LOF, NND and LNND, and LOF and NND outperform LNND. The performance of ALP and SVM is comparable, but ALP can be optimised more efficiently so constitutes a good default choice. Alternatively, using validation AUROC as a selection criterion between ALP or SVM gives the best overall result, and NND is the least computationally demanding option. We thus end up with a clear trade-off between three choices, allowing practitioners to make an informed decision.
引用
收藏
页码:2863 / 2883
页数:20
相关论文
共 69 条
[1]  
Bekker J(2020)Learning from positive and unlabeled data: A survey Machine Learning 109 719-760
[2]  
Davis J(2016)Should we really use post-hoc tests based on mean-ranks? Journal of Machine Learning Research 17 152-161
[3]  
Benavoli A(2012)Random search for hyper-parameter optimization Journal of Machine Learning Research 13 281-305
[4]  
Corani G(1991)Bayesian methods in global optimization Journal of Global Optimization 1 1-14
[5]  
Mangili F(2020)Pymoo: Multi-objective optimization in Python IEEE Access 8 89497-89509
[6]  
Bergstra J(1995)Support-vector networks Machine Learning 20 273-297
[7]  
Bengio Y(2006)Statistical comparisons of classifiers over multiple data sets Journal of Machine Learning Research 7 1-30
[8]  
Betrò B(2018)Two combination stages of clustered one-class classifiers for writer identification from text fragments Pattern Recognition 82 147-162
[9]  
Blank J(1979)A simple sequentially rejective multiple test procedure Scandinavian Journal of Statistics 6 65-70
[10]  
Deb K(1961)"Direct search" solution of numerical and statistical problems Journal of the ACM 8 212-229