Error estimation based on variance analysis of k-fold cross-validation

被引:120
作者
Jiang, Gaoxia [1 ]
Wang, Wenjian [1 ]
机构
[1] Shanxi Univ, Sch Comp & Informat Technol, Taiyuan 030006, Peoples R China
基金
中国国家自然科学基金;
关键词
Error estimation; k-fold cross-validation; Variance analysis; Model selection; ALGORITHMS;
D O I
10.1016/j.patcog.2017.03.025
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-validation (CV) is often used to estimate the generalization capability of a learning model. The variance of CV error has a considerable impact on the accuracy of CV estimator and the adequacy of the learning model, so it is very important to analyze CV variance. The aim of this paper is to investigate how to improve the accuracy of the error estimation based on variance analysis. We first describe the quantitative relationship between CV variance and its accuracy, which can provide guidance for improving the accuracy by reducing the variance. We then study the relationships between variance and relevant variables including the sample size, the number of folds, and the number of repetitions. These form the basis of theoretical strategies of regulating CV variance. Our classification results can theoretically explain the empirical results of Rodriguez and Kohavi. Finally, we propose a uniform normalized variance which not only measures model accuracy but also is irrelative to fold number. Therefore, it simplifies the selection of fold number in k-fold CV and normalized variance can serve as a stable error measurement for model comparison and selection. We report the results of experiments using 5 supervised learning models and 20 datasets. The results indicate that it is reliable to determine which variance is less before k-fold CV by the proposed theorems, and thus the accuracy of error estimation can be promoted by reducing variance. In so doing, we are more likely to select the best parameter or model. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:94 / 106
页数:13
相关论文
共 30 条
[1]   Combined 5 x 2 cv F test for comparing supervised classification learning algorithms [J].
Alpaydin, E .
NEURAL COMPUTATION, 1999, 11 (08) :1885-1892
[2]   Fast cross-validation algorithms for least squares support vector machine and kernel ridge regression [J].
An, Senjian ;
Liu, Wanquan ;
Venkatesh, Svetha .
PATTERN RECOGNITION, 2007, 40 (08) :2154-2162
[3]  
[Anonymous], 1995, THESIS STANFORD U
[4]  
[Anonymous], INTRO BOOTSTRAP
[5]  
Bengio Y, 2004, J MACH LEARN RES, V5, P1089
[6]   On the use of cross-validation for time series predictor evaluation [J].
Bergmeir, Christoph ;
Benitez, Jose M. .
INFORMATION SCIENCES, 2012, 191 :192-213
[7]  
Bergstra J, 2012, J MACH LEARN RES, V13, P281
[8]  
Chang C.-C., 2015, LIBSVM DATA CLASSIFI
[9]  
DeGroot M. H., 2011, PROBABILITY STAT, P337
[10]   Sensitivity Analysis of k-Fold Cross Validation in Prediction Error Estimation [J].
Diego Rodriguez, Juan ;
Perez, Aritz ;
Antonio Lozano, Jose .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2010, 32 (03) :569-575