A Study on the Relationships of Classifier Performance Metrics

被引:149
作者
Seliya, Naeem [1 ]
Khoshgoftaar, Taghi M. [2 ]
Van Hulse, Jason [2 ]
机构
[1] Univ Michigan, Comp & Informat Sci, 4901 Evergreen Rd, Dearborn, MI 48128 USA
[2] Florida Atlantic Univ, Comp Sci & Engn, Boca Raton, FL 33431 USA
来源
ICTAI: 2009 21ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE | 2009年
关键词
binary classification; performance metrics; factor analysis; metrics relationship;
D O I
10.1109/ICTAI.2009.25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is no general consensus on which classifier performance metrics are better to use as compared to others. While some studies investigate a handful of such metrics in a comparative fashion, an evaluation of specific relationships among a large set of commonly-used performance metrics is much needed in the data mining and machine learning community. This study provides a unique insight into the underlying relationships among classifier performance metrics. We do so with a large case study involving 35 datasets from various domains and the C4.5 decision tree algorithm. A common property of the 35 datasets is that they suffer from the class imbalance problem. Our approach is based on applying factor analysis to the classifier performance space which is characterized by 22 performance metrics. It is shown that such a large number of performance metrics can be grouped into two-to-four relationship-based groups extracted by factor analysis. This work is a step in the direction of providing the analyst with an improved understanding about the different relationships and groupings among the performance metrics, thus facilitating the selection of performance metrics that capture relatively independent aspects of a classifier's performance.
引用
收藏
页码:59 / +
页数:2
相关论文
共 20 条
[1]   Visualizing Classifier Performance on Different Domains [J].
Alaiz-Rodriguez, Rocio ;
Japkowicz, Nathalie ;
Tischer, Peter .
20TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL 2, PROCEEDINGS, 2008, :3-+
[2]  
[Anonymous], PATTERN RECOGNITION
[3]  
[Anonymous], 1988, Applied Multivariate Statistical Analysis
[4]  
Asuncion A., 2007, UCI Machine Learning Repository
[5]  
CARUANA R, 2004, P 10 ACM INT C KNOWL
[6]  
Davis J., 2006, P 23 INT C MACH LEAR, P115
[7]  
HAND DJ, 1997, J ROYAL STAT SOC A, P521
[8]  
Hoadley B., 1998, IMA Journal of Mathematics Applied in Business and Industry, V9, P55, DOI 10.1093/imaman/9.1.55
[9]  
Japkowicz N., 2008, P 3 WORKSH EV METH M
[10]   Comparative assessment of software quality classification techniques: An empirical case study [J].
Khoshgoftaar, TM ;
Seliya, N .
EMPIRICAL SOFTWARE ENGINEERING, 2004, 9 (03) :229-257