Measuring classifier performance: a coherent alternative to the area under the ROC curve

被引:0
作者
David J. Hand
机构
[1] Imperial College London,Department of Mathematics
[2] Imperial College London,Institute for Mathematical Sciences
来源
Machine Learning | 2009年 / 77卷
关键词
ROC curves; Classification; Specificity; Sensitivity; Misclassification rate; Cost; Loss; Error rate;
D O I
暂无
中图分类号
学科分类号
摘要
The area under the ROC curve (AUC) is a very widely used measure of performance for classification and diagnostic rules. It has the appealing property of being objective, requiring no subjective input from the user. On the other hand, the AUC has disadvantages, some of which are well known. For example, the AUC can give potentially misleading results if ROC curves cross. However, the AUC also has a much more serious deficiency, and one which appears not to have been previously recognised. This is that it is fundamentally incoherent in terms of misclassification costs: the AUC uses different misclassification cost distributions for different classifiers. This means that using the AUC is equivalent to using different metrics to evaluate different classification rules. It is equivalent to saying that, using one classifier, misclassifying a class 1 point is p times as serious as misclassifying a class 0 point, but, using another classifier, misclassifying a class 1 point is P times as serious, where p≠P. This is nonsensical because the relative severities of different kinds of misclassifications of individual points is a property of the problem, not the classifiers which happen to have been chosen. This property is explored in detail, and a simple valid alternative to the AUC is proposed.
引用
收藏
页码:103 / 123
页数:20
相关论文
共 50 条
[31]   Making classifier performance comparisons when ROC curves intersect [J].
Gigliarano, Chiara ;
Figini, Silvia ;
Muliere, Pietro .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2014, 77 :300-312
[32]   FINDING THE OPTIMAL THRESHOLD OF A PARAMETRIC ROC CURVE UNDER A CONTINUOUS DIAGNOSTIC MEASUREMENT [J].
Hwang, Yi-Ting ;
Hung, Yu-Han ;
Wang, Chun Chao ;
Terng, Harn-Jing .
REVSTAT-STATISTICAL JOURNAL, 2018, 16 (01) :23-43
[33]   Comparing Classifier's Performance Based on Confidence Interval of the ROC [J].
Malach, Tobias ;
Pomenkova, Jitka .
RADIOENGINEERING, 2018, 27 (03) :827-834
[34]   FUNCTIONAL COVARIATE-ADJUSTED PARTIAL AREA UNDER THE SPECIFICITY-ROC CURVE WITH AN APPLICATION TO METABOLIC SYNDROME DIAGNOSIS [J].
Inacio de Carvalho, Vanda ;
de Carvalho, Miguel ;
Alonzo, Todd A. ;
Gonzalez-Manteiga, Wenceslao .
ANNALS OF APPLIED STATISTICS, 2016, 10 (03) :1472-1495
[35]   The linear combinations of biomarkers which maximize the partial area under the ROC curves [J].
Man-Jen Hsu ;
Huey-Miin Hsueh .
Computational Statistics, 2013, 28 :647-666
[36]   The linear combinations of biomarkers which maximize the partial area under the ROC curves [J].
Hsu, Man-Jen ;
Hsueh, Huey-Miin .
COMPUTATIONAL STATISTICS, 2013, 28 (02) :647-666
[37]   cvauroc: Command to compute cross-validated area under the curve for ROC analysis after predictive modeling for binary outcomes [J].
Angel Luque-Fernandez, Miguel ;
Redondo-Sanchez, Daniel ;
Maringe, Camille .
STATA JOURNAL, 2019, 19 (03) :615-625
[38]   Comparing Multi-objective and Threshold-moving ROC Curve Generation for a Prototype-based Classifier [J].
Aler, Ricardo ;
Handl, Julia ;
Knowles, Joshua D. .
GECCO'13: PROCEEDINGS OF THE 2013 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2013, :1029-1036
[39]   Grouped Variable Selection Using Area under the ROC with Imbalanced Data [J].
Li, Yang ;
Qin, Yichen ;
Wang, Limin ;
Chen, Jiaxu ;
Ma, Shuangge .
COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2016, 45 (04) :1268-1280
[40]   The clinical meaning of the area under a receiver operating characteristic curve for the evaluation of the performance of disease markers [J].
Parodi, Stefano ;
Verda, Damiano ;
Bagnasco, Francesca ;
Muselli, Marco .
EPIDEMIOLOGY AND HEALTH, 2022, 44