Evaluating Classification Model Against Bayes Error Rate

被引:6
作者
Chen, Qingqiang [1 ]
Cao, Fuyuan [1 ]
Xing, Ying [1 ]
Liang, Jiye [1 ]
机构
[1] Shanxi Univ, Sch Comp & Informat Technol, Key Lab Computat Intelligence & Chinese Informat, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Bayes error rate; classification; label propagation; model evaluation; percolation theory; LOWER BOUNDS; TIGHT UPPER; PROBABILITY; DIVERGENCE; DISTANCE;
D O I
10.1109/TPAMI.2023.3240194
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For a classification task, we usually select an appropriate classifier via model selection. How to evaluate whether the chosen classifier is optimal? One can answer this question via Bayes error rate (BER). Unfortunately, estimating BER is a fundamental conundrum. Most existing BER estimators focus on giving the upper and lower bounds of the BER. However, evaluating whether the selected classifier is optimal based on these bounds is hard. In this article, we aim to learn the exact BER instead of bounds on BER. The core of our method is to transform the BER calculation problem into a noise recognition problem. Specifically, we define a type of noise called Bayes noise and prove that the proportion of Bayes noisy samples in a data set is statistically consistent with the BER of the data set. To recognize the Bayes noisy samples, we present a method consisting of two parts: selecting reliable samples based on percolation theory and then employing a label propagation algorithm to recognize the Bayes noisy samples based on the selected reliable samples. The superiority of the proposed method compared to the existing BER estimators is verified on extensive synthetic, benchmark, and image data sets.
引用
收藏
页码:9639 / 9653
页数:15
相关论文
共 51 条
[1]  
Alcalá-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
[2]   Lower bounds for Bayes error estimation [J].
Antos, A ;
Devroye, L ;
Györfi, L .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1999, 21 (07) :643-645
[3]   Arbitrarily tight upper and lower bounds on the Bayesian probability of error [J].
AviItzhak, H ;
Diep, T .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1996, 18 (01) :89-91
[4]   Empirically Estimable Classification Bounds Based on a Nonparametric Divergence Measure [J].
Berisha, Visar ;
Wisler, Alan ;
Hero, Alfred O., III ;
Spanias, Andreas .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (03) :580-591
[5]  
Bossard L, 2014, LECT NOTES COMPUT SC, V8694, P446, DOI 10.1007/978-3-319-10599-4_29
[6]  
Breiman L, 2001, MACH LEARN, V45, P5, DOI [10.1186/s12859-018-2419-4, 10.3322/caac.21834]
[7]   LIBSVM: A Library for Support Vector Machines [J].
Chang, Chih-Chung ;
Lin, Chih-Jen .
ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
[8]  
Coates A., 2011, AISTATS
[9]  
Cover T. M, 1969, LEARNING PATTERN REC
[10]   NEAREST NEIGHBOR PATTERN CLASSIFICATION [J].
COVER, TM ;
HART, PE .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) :21-+