An efficient method for feature selection in linear regression based on an extended Akaike's information criterion

被引:1
|
作者
Vetrov, D. P. [1 ]
Kropotov, D. A. [2 ]
Ptashko, N. O. [1 ]
机构
[1] Moscow MV Lomonosov State Univ, Fac Computat Math & Cybernet, Moscow 119992, Russia
[2] Russian Acad Sci, Dorodnicyn Comp Ctr, Moscow 119333, Russia
基金
俄罗斯基础研究基金会;
关键词
pattern recognition; linear regression; feature selection; Akaike's information criterion;
D O I
10.1134/S096554250911013X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
A method for feature selection in linear regression based on an extension of Akaike's information criterion is proposed. The use of classical Akaike's information criterion (AIC) for feature selection assumes the exhaustive search through all the subsets of features, which has unreasonably high computational and time cost. A new information criterion is proposed that is a continuous extension of AIC. As a result, the feature selection problem is reduced to a smooth optimization problem. An efficient procedure for solving this problem is derived. Experiments show that the proposed method enables one to efficiently select features in linear regression. In the experiments, the proposed procedure is compared with the relevance vector machine, which is a feature selection method based on Bayesian approach. It is shown that both procedures yield similar results. The main distinction of the proposed method is that certain regularization coefficients are identical zeros. This makes it possible to avoid the underfitting effect, which is a characteristic feature of the relevance vector machine. A special case (the so-called nondiagonal regularization) is considered in which both methods are identical.
引用
收藏
页码:1972 / 1985
页数:14
相关论文
共 50 条
  • [21] Scale-Invariant and consistent Bayesian information criterion for order selection in linear regression models
    Gohain, Prakash B.
    Jansson, Magnus
    SIGNAL PROCESSING, 2022, 196
  • [22] Dynamic feature selection method with minimum redundancy information for linear data
    HongFang Zhou
    Jing Wen
    Applied Intelligence, 2020, 50 : 3660 - 3677
  • [23] Dynamic feature selection method with minimum redundancy information for linear data
    Zhou, HongFang
    Wen, Jing
    APPLIED INTELLIGENCE, 2020, 50 (11) : 3660 - 3677
  • [24] An efficient bit-based feature selection method
    Chen, Wei-Chou
    Tseng, Shian-Shyong
    Hong, Tzung-Pei
    EXPERT SYSTEMS WITH APPLICATIONS, 2008, 34 (04) : 2858 - 2869
  • [25] A novel permission-based Android malware detection system using feature selection based on linear regression
    Durmuş Özkan Şahin
    Oğuz Emre Kural
    Sedat Akleylek
    Erdal Kılıç
    Neural Computing and Applications, 2023, 35 : 4903 - 4918
  • [26] A novel permission-based Android malware detection system using feature selection based on linear regression
    Sahin, Durmus Ozkan
    Kural, Oguz Emre
    Akleylek, Sedat
    Kilic, Erdal
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (07) : 4903 - 4918
  • [27] Linear feature selection in texture analysis - A PLS based method
    Joselene Marques
    Christian Igel
    Martin Lillholm
    Erik B. Dam
    Machine Vision and Applications, 2013, 24 : 1435 - 1444
  • [28] Linear feature selection in texture analysis - A PLS based method
    Marques, Joselene
    Igel, Christian
    Lillholm, Martin
    Dam, Erik B.
    MACHINE VISION AND APPLICATIONS, 2013, 24 (07) : 1435 - 1444
  • [29] Feature selection method based on fuzzy entropy for regression in QSAR studies
    Elmi, Zahra
    Faez, Karim
    Goodarzi, Mohammad
    Goudarzi, Nasser
    MOLECULAR PHYSICS, 2009, 107 (17) : 1787 - 1798
  • [30] Quick online feature selection method for regression - A feature selection method inspired by human behavior
    Tadeuchi, Youhei
    Oshima, Ryuji
    Nishida, Kyosuke
    Yamauchi, Koichiro
    Omori, Takashi
    2007 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-8, 2007, : 3227 - +