Information criteria: How do they behave in different models?

被引:61
作者
Emiliano, Paulo C. [1 ]
Vivanco, Mario J. F. [1 ]
de Menezes, Fortunato S. [1 ]
机构
[1] Fed Univ Lavras UFLA, Dept Exact Sci DEX, BR-37200000 Lavras, MG, Brazil
关键词
Akaike information criterion (AIC); Entropy; Schwarz information criterion; BIC; Kullback-Leibler information; Selection of models; GROWTH-CURVES; TIME-SERIES; REGRESSION; EFFICIENCY; SELECTION; MULTIPLE;
D O I
10.1016/j.csda.2013.07.032
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The choice of the best model is crucial in modeling data, and parsimony is one of the principles that must guide this choice. Despite their broad use in model selection, the foundations of the Akaike information criterion (AIC), the corrected Akaike criterion (AICc) and the Bayesian information criterion (BIC) are, in general, poorly understood. The AIC, AICc and BIC penalize the likelihoods in order to select the simplest model. These criteria are based upon concepts of information and entropy, which are explained in this work, by focusing on a statistical approach. The three criteria are compared through Monte Carlo simulations, and the applications of these criteria are investigated in the selection of normal models, the selection of biological growth models and selection of time series models. For the simulation with normal models, all three criteria exhibited poor performance for a small sample size N = 100 (particularly, when the variances are slightly different). For biological growth model simulations with a very small sample size N = 13 the AIC and AICc showed better performance in comparison to the BIC. The simulation based on time series models produced results similar to the normal model simulations. For these simulations, the BIC exhibited superior performance, in some cases, in comparison to the other two information criteria (AIC and AICc) for a small sample size N = 100, but in other cases, the BIC performed poorly, as did the AIC and AICc. (C) 2013 Elsevier B.V. All rights reserved.
引用
收藏
页码:141 / 153
页数:13
相关论文
共 36 条
  • [1] NEW LOOK AT STATISTICAL-MODEL IDENTIFICATION
    AKAIKE, H
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) : 716 - 723
  • [2] [Anonymous], 2001, The elements of statistical learning: data mining, inference and prediction
  • [3] [Anonymous], 2006, Pattern recognition and machine learning
  • [4] Brody Samuel, 1945
  • [5] Multimodel inference - understanding AIC and BIC in model selection
    Burnham, KP
    Anderson, DR
    [J]. SOCIOLOGICAL METHODS & RESEARCH, 2004, 33 (02) : 261 - 304
  • [6] Burnham KP., 2002, MODEL SELECTION MULT, DOI DOI 10.1007/B97636
  • [7] Goodness of fit via non-parametric likelihood ratios
    Claeskens, G
    Hjort, NL
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 2004, 31 (04) : 487 - 513
  • [8] Claeskens G., 2008, Model Selection and Model Averaging, DOI DOI 10.1017/CBO9780511790485
  • [9] GENETIC AND ENVIRONMENTAL ASPECTS OF THE GROWTH CURVE PARAMETERS IN BEEF-COWS
    DENISE, RSK
    BRINKS, JS
    [J]. JOURNAL OF ANIMAL SCIENCE, 1985, 61 (06) : 1431 - 1440
  • [10] Draper N. R., 1998, APPL REGRESSION ANAL, DOI DOI 10.1002/9781118625590.CH15