An efficient method for feature selection in linear regression based on an extended Akaike's information criterion

被引:1
|
作者
Vetrov, D. P. [1 ]
Kropotov, D. A. [2 ]
Ptashko, N. O. [1 ]
机构
[1] Moscow MV Lomonosov State Univ, Fac Computat Math & Cybernet, Moscow 119992, Russia
[2] Russian Acad Sci, Dorodnicyn Comp Ctr, Moscow 119333, Russia
基金
俄罗斯基础研究基金会;
关键词
pattern recognition; linear regression; feature selection; Akaike's information criterion;
D O I
10.1134/S096554250911013X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
A method for feature selection in linear regression based on an extension of Akaike's information criterion is proposed. The use of classical Akaike's information criterion (AIC) for feature selection assumes the exhaustive search through all the subsets of features, which has unreasonably high computational and time cost. A new information criterion is proposed that is a continuous extension of AIC. As a result, the feature selection problem is reduced to a smooth optimization problem. An efficient procedure for solving this problem is derived. Experiments show that the proposed method enables one to efficiently select features in linear regression. In the experiments, the proposed procedure is compared with the relevance vector machine, which is a feature selection method based on Bayesian approach. It is shown that both procedures yield similar results. The main distinction of the proposed method is that certain regularization coefficients are identical zeros. This makes it possible to avoid the underfitting effect, which is a characteristic feature of the relevance vector machine. A special case (the so-called nondiagonal regularization) is considered in which both methods are identical.
引用
收藏
页码:1972 / 1985
页数:14
相关论文
共 50 条
  • [11] A primer on model selection using the Akaike Information Criterion
    Portet, Stephanie
    INFECTIOUS DISEASE MODELLING, 2020, 5 : 111 - 128
  • [12] On bias correction of the Akaike information criterion in linear models
    Noda, K
    Miyaoka, E
    Itoh, M
    COMMUNICATIONS IN STATISTICS-THEORY AND METHODS, 1996, 25 (08) : 1845 - 1857
  • [13] APPLICATION OF AKAIKE INFORMATION CRITERION STATISTICS TO GEOTECHNICAL INVERSE ANALYSIS - THE EXTENDED BAYESIAN METHOD
    HONJO, Y
    LIU, WT
    SAKAJO, S
    STRUCTURAL SAFETY, 1994, 14 (1-2) : 5 - 29
  • [14] A new segmentation method of electroencephalograms by use of Akaike's information criterion
    Inouye, T
    Toi, S
    Matsumoto, Y
    COGNITIVE BRAIN RESEARCH, 1995, 3 (01): : 33 - 40
  • [15] Hybrid Feature Selection: Combining Fisher Criterion and Mutual Information for Efficient Feature Selection
    Dhir, Chandra Shekhar
    Lee, Soo Young
    ADVANCES IN NEURO-INFORMATION PROCESSING, PT I, 2009, 5506 : 613 - 620
  • [16] Bootstrap variants of the Akaike information criterion for mixed model selection
    Shang, Junfeng
    Cavanaugh, Joseph E.
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2008, 52 (04) : 2004 - 2021
  • [17] A general Akaike-type criterion for model selection in robust regression
    Burman, P
    Nolan, D
    BIOMETRIKA, 1995, 82 (04) : 877 - 886
  • [18] Comparison of Akaike information criterion and consistent Akaike information criterion for model selection and statistical inference from capture-recapture studies
    Anderson, DR
    Burnham, KP
    White, GC
    JOURNAL OF APPLIED STATISTICS, 1998, 25 (02) : 263 - 282
  • [19] Evaluation of Five Classifiers for Children Activity Recognition with Sound as Information Source and Akaike Criterion for Feature Selection
    Garcia-Dominguez, Antonio
    Zanella-Calzada, Laura A.
    Galvan-Tejada, Carlos E.
    Galvan-Tejada, Jorge I.
    Celaya-Padilla, Jose M.
    PATTERN RECOGNITION, MCPR 2019, 2019, 11524 : 398 - 407
  • [20] Asymptotic post-selection inference for the Akaike information criterion
    Charkhi, Ali
    Claeskens, Gerda
    BIOMETRIKA, 2018, 105 (03) : 645 - 664