On Over-fitting in Model Selection and Subsequent Selection Bias in Performance Evaluation

被引:0
|
作者
Cawley, Gavin C. [1 ]
Talbot, Nicola L. C. [1 ]
机构
[1] Univ E Anglia, Sch Comp Sci, Norwich NR4 7TJ, Norfolk, England
基金
英国工程与自然科学研究理事会;
关键词
model selection; performance evaluation; bias-variance trade-off; selection bias; over-fitting; SUPPORT VECTOR MACHINE; OUT CROSS-VALIDATION; CLASSIFICATION; REGULARIZATION; COEFFICIENTS; PARAMETERS; STABILITY; VARIANCE; NETWORKS; BOUNDS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Model selection strategies for machine learning algorithms typically involve the numerical optimisation of an appropriate model selection criterion, often based on an estimator of generalisation performance, such as k-fold cross-validation. The error of such an estimator can be broken down into bias and variance components. While unbiasedness is often cited as a beneficial quality of a model selection criterion, we demonstrate that a low variance is at least as important, as a non-negligible variance introduces the potential for over-fitting in model selection as well as in training the model. While this observation is in hindsight perhaps rather obvious, the degradation in performance due to over-fitting the model selection criterion can be surprisingly large, an observation that appears to have received little attention in the machine learning literature to date. In this paper, we show that the effects of this form of over-fitting are often of comparable magnitude to differences in performance between learning algorithms, and thus cannot be ignored in empirical evaluation. Furthermore, we show that some common performance evaluation practices are susceptible to a form of selection bias as a result of this form of over-fitting and hence are unreliable. We discuss methods to avoid over-fitting in model selection and subsequent selection bias in performance evaluation, which we hope will be incorporated into best practice. While this study concentrates on cross-validation based model selection, the findings are quite general and apply to any model selection practice involving the optimisation of a model selection criterion evaluated over a finite sample of data, including maximisation of the Bayesian evidence and optimisation of performance bounds.
引用
收藏
页码:2079 / 2107
页数:29
相关论文
共 50 条
  • [1] Preventing over-fitting during model selection via Bayesian regularisation of the hyper-parameters
    Cawley, Gavin C.
    Talbot, Nicola L. C.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2007, 8 : 841 - 861
  • [2] A modification of the bootstrapping soft shrinkage approach for spectral variable selection in the issue of over-fitting, model accuracy and variable selection credibility
    Yan, Hong
    Song, Xiangzhong
    Tian, Kuangda
    Gao, Jingxian
    Li, Qianqian
    Xiong, Yanmei
    Min, Shungeng
    SPECTROCHIMICA ACTA PART A-MOLECULAR AND BIOMOLECULAR SPECTROSCOPY, 2019, 210 : 362 - 371
  • [3] Handling over-fitting in test cost-sensitive decision tree learning by feature selection, smoothing and pruning
    Wang, Tao
    Qin, Zhenxing
    Jin, Zhi
    Zhang, Shichao
    JOURNAL OF SYSTEMS AND SOFTWARE, 2010, 83 (07) : 1137 - 1147
  • [4] A new strategy to prevent over-fitting in partial least squares models based on model population analysis
    Deng, Bai-Chuan
    Yun, Yong-Huan
    Liang, Yi-Zeng
    Cao, Dong-Sheng
    Xu, Qing-Song
    Yi, Lun-Zhao
    Huang, Xin
    ANALYTICA CHIMICA ACTA, 2015, 880 : 32 - 41
  • [5] A General Evaluation System for Optimal Selection Performance of Radar Clutter Model
    Yang, Wei
    Zhang, Liang
    Yang, Liru
    Zhang, Wenpeng
    Shen, Qingmu
    JOURNAL OF SYSTEMS ENGINEERING AND ELECTRONICS, 2023, 34 (06) : 1520 - 1525
  • [6] Study of the Over-Fitting in Building PLS Model Using Orthogonal Signal Correction
    Zhang Xian
    Yuan Hong-fu
    Guo Zheng
    Song Chun-feng
    Li Xiao-yu
    Xie Jin-chun
    SPECTROSCOPY AND SPECTRAL ANALYSIS, 2011, 31 (06) : 1688 - 1691
  • [7] RESAMPLING METHODS FOR MODEL FITTING AND MODEL SELECTION
    Babu, G. Jogesh
    JOURNAL OF BIOPHARMACEUTICAL STATISTICS, 2011, 21 (06) : 1177 - 1186
  • [8] Impact of Selection Bias on Estimation of Subsequent Event Risk
    Hu, Yi-Juan
    Schmidt, Amand F.
    Dudbridg, Frank
    Holmes, Michael V.
    Brophy, James M.
    Tragante, Vinicius
    Li, Ziyi
    Liao, Peizhou
    Quyyumi, Arshed A.
    McCubreu, Raymond O.
    Horne, Benjamin D.
    Hingorani, Aroon D.
    Asselbergs, Folkert W.
    Patel, Riyaz S.
    Long, Qi
    CIRCULATION-CARDIOVASCULAR GENETICS, 2017, 10 (05)
  • [9] High-Dimensional Data and the Bias Variance Tradeoff in Model Selection
    Menna, Eligo Workineh
    INTERNATIONAL JOURNAL OF APPLIED MATHEMATICS & STATISTICS, 2024, 63 : 34 - 56
  • [10] Selection bias in credit scorecard evaluation
    Hand, David J.
    Adams, Niall M.
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 2014, 65 (03) : 408 - 415