An efficient method for feature selection in linear regression based on an extended Akaike's information criterion

被引:1
|
作者
Vetrov, D. P. [1 ]
Kropotov, D. A. [2 ]
Ptashko, N. O. [1 ]
机构
[1] Moscow MV Lomonosov State Univ, Fac Computat Math & Cybernet, Moscow 119992, Russia
[2] Russian Acad Sci, Dorodnicyn Comp Ctr, Moscow 119333, Russia
基金
俄罗斯基础研究基金会;
关键词
pattern recognition; linear regression; feature selection; Akaike's information criterion;
D O I
10.1134/S096554250911013X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
A method for feature selection in linear regression based on an extension of Akaike's information criterion is proposed. The use of classical Akaike's information criterion (AIC) for feature selection assumes the exhaustive search through all the subsets of features, which has unreasonably high computational and time cost. A new information criterion is proposed that is a continuous extension of AIC. As a result, the feature selection problem is reduced to a smooth optimization problem. An efficient procedure for solving this problem is derived. Experiments show that the proposed method enables one to efficiently select features in linear regression. In the experiments, the proposed procedure is compared with the relevance vector machine, which is a feature selection method based on Bayesian approach. It is shown that both procedures yield similar results. The main distinction of the proposed method is that certain regularization coefficients are identical zeros. This makes it possible to avoid the underfitting effect, which is a characteristic feature of the relevance vector machine. A special case (the so-called nondiagonal regularization) is considered in which both methods are identical.
引用
收藏
页码:1972 / 1985
页数:14
相关论文
共 50 条
  • [11] ROBUST CRITERION FOR VARIABLE SELECTION IN LINEAR REGRESSION
    Patil, A. B.
    Kashid, D. N.
    INTERNATIONAL JOURNAL OF AGRICULTURAL AND STATISTICAL SCIENCES, 2009, 5 (02): : 509 - 521
  • [12] Mixed Integer Nonlinear Program for Minimization of Akaike's Information Criterion
    Kimura, Keiji
    Waki, Hayato
    MATHEMATICAL SOFTWARE, ICMS 2016, 2016, 9725 : 292 - 300
  • [13] Feature Selection Method with Feature Subcategorization in Regression
    Wanwan Zheng
    SN Computer Science, 6 (5)
  • [14] A brief guide to model selection, multimodel inference and model averaging in behavioural ecology using Akaike’s information criterion
    Matthew R. E. Symonds
    Adnan Moussalli
    Behavioral Ecology and Sociobiology, 2011, 65 : 13 - 21
  • [15] A brief guide to model selection, multimodel inference and model averaging in behavioural ecology using Akaike's information criterion
    Symonds, Matthew R. E.
    Moussalli, Adnan
    BEHAVIORAL ECOLOGY AND SOCIOBIOLOGY, 2011, 65 (01) : 13 - 21
  • [16] A Robust Linear Regression Feature Selection Method for Data Sets With Unknown Noise
    Guo, Yaqing
    Wang, Wenjian
    Wang, Xuejun
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (01) : 31 - 44
  • [17] Is mutual information adequate for feature selection in regression?
    Frenay, Benoit
    Doquire, Gauthier
    Verleysen, Michel
    NEURAL NETWORKS, 2013, 48 : 1 - 7
  • [18] Robust Information Criterion for Model Selection in Sparse High-Dimensional Linear Regression Models
    Gohain, Prakash Borpatra
    Jansson, Magnus
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 2251 - 2266
  • [19] Multilabel Feature Selection Based on Fuzzy Mutual Information and Orthogonal Regression
    Dai, Jianhua
    Liu, Qi
    Chen, Wenxiang
    Zhang, Chucai
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (09) : 5136 - 5148
  • [20] Feature Selection with Mutual Information for Regression Problems
    Sulaiman, Muhammad Aliyu
    Labadin, Jane
    2015 9TH INTERNATIONAL CONFERENCE ON IT IN ASIA (CITA), 2015,