Generalization Performance of Pure Accuracy and its Application in Selective Ensemble Learning

被引:12
作者
Wang, Jieting [1 ]
Qian, Yuhua [1 ,2 ]
Li, Feijiang [1 ]
Liang, Jiye [2 ]
Zhang, Qingfu [3 ]
机构
[1] Shanxi Univ, Inst Big Data Sci & Ind, Taiyuan 030006, Shanxi, Peoples R China
[2] Shanxi Univ, Key Lab Computat Intelligence & Chinese Informat P, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
[3] City Univ Hong Kong, Dept Comp Sci, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Generalization performance bound; linear-fractional measure; pure accuracy; selective ensemble learning; CONCENTRATION INEQUALITY; RISK; CLASSIFICATION; RELIABILITY; BOUNDS;
D O I
10.1109/TPAMI.2022.3171436
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
pure accuracy measure is used to eliminate random consistency from the accuracy measure. Biases to both majority and minority classes in the pure accuracy are lower than that in the accuracy measure. In this paper, we demonstrate that compared with the accuracy measure and F-measure, the pure accuracy measure is class distribution insensitive and discriminative for good classifiers. The advantages make the pure accuracy measure suitable for traditional classification. Further, we mainly focus on two points: exploring a tighter generalization bound on pure accuracy based learning paradigm and designing a learning algorithm based on the pure accuracy measure. Particularly, with the self-bounding property, we build an algorithm-independent generalization bound on the pure accuracy measure, which is tighter than the existing bound of an order O(1/root N ) (N is the number of instances). The proposed bound is free from making a smoothness or convex assumption on the hypothesis functions. In addition, we design a learning algorithm optimizing the pure accuracy measure and use it in the selective ensemble learning setting. The experiments on sixteen benchmark data sets and four image data sets demonstrate that the proposed method statistically performs better than the other eight representative benchmark algorithms.
引用
收藏
页码:1798 / 1816
页数:19
相关论文
共 65 条
[1]  
Agarwal S, 2005, J MACH LEARN RES, V6, P393
[2]   The complexity of exact learning of acyclic conditional preference networks from swap examples [J].
Alanazi, Eisa ;
Mouhoub, Malek ;
Zilles, Sandra .
ARTIFICIAL INTELLIGENCE, 2020, 278
[3]   On similarity indices and correction for chance agreement [J].
Albatineh, Ahmed N. ;
Niewiadomska-Bugaj, Magdalena ;
Mihalko, Daniel .
JOURNAL OF CLASSIFICATION, 2006, 23 (02) :301-313
[4]   KEEL: a software tool to assess evolutionary algorithms for data mining problems [J].
Alcala-Fdez, J. ;
Sanchez, L. ;
Garcia, S. ;
del Jesus, M. J. ;
Ventura, S. ;
Garrell, J. M. ;
Otero, J. ;
Romero, C. ;
Bacardit, J. ;
Rivas, V. M. ;
Fernandez, J. C. ;
Herrera, F. .
SOFT COMPUTING, 2009, 13 (03) :307-318
[5]  
Bartlett P. L., 2003, Journal of Machine Learning Research, V3, P463, DOI 10.1162/153244303321897690
[6]   Convexity, classification, and risk bounds [J].
Bartlett, PL ;
Jordan, MI ;
McAuliffe, JD .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2006, 101 (473) :138-156
[7]  
Boucheron S, 2000, RANDOM STRUCT ALGOR, V16, P277, DOI 10.1002/(SICI)1098-2418(200005)16:3<277::AID-RSA4>3.0.CO
[8]  
2-1
[9]  
Boucheron S., 2013, CONCENTRATION INEQUA, DOI 10.1093/acprof:oso/9780199535255.001.0001
[10]   A Bennett concentration inequality and its application to suprema of empirical processes [J].
Bousquet, O .
COMPTES RENDUS MATHEMATIQUE, 2002, 334 (06) :495-500