Ensemble Feature Selection With Block-Regularized m x 2 Cross-Validation

被引:2
|
作者
Yang, Xingli [1 ]
Wang, Yu [2 ]
Wang, Ruibo [2 ]
Li, Jihong [2 ]
机构
[1] Shanxi Univ, Sch Math Sci, Taiyuan 030006, Peoples R China
[2] Shanxi Univ, Sch Modern Educ Technol, Taiyuan 030006, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Correlation; Indexes; Data models; Technological innovation; Reliability theory; Upper bound; Beta distribution; block-regularized m x 2 cross-validation; ensemble feature selection (EFS); false positive; true positive; VARIABLE SELECTION; REGRESSION; PRECISION; RECALL;
D O I
10.1109/TNNLS.2021.3128173
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Ensemble feature selection (EFS) has attracted significant interest in the literature due to its great potential in reducing the discovery rate of noise features and stabilizing the feature selection results. In view of the superior performance of block-regularized m x 2 cross-validation on generalization performance and algorithm comparison, a novel EFS technology based on block-regularized m x 2 cross-validation is proposed in this study. Contrary to the traditional ensemble learning with a binomial distribution, the distribution of feature selection frequency in the proposed technique is approximated by a beta distribution more accurately. Furthermore, theoretical analysis of the proposed technique shows that it yields a higher selection probability for important features, lower selected risk for noise features, more true positives, and fewer false positives. Finally, the above conclusions are verified by the simulated and real data experiments.
引用
收藏
页码:6628 / 6641
页数:14
相关论文
共 50 条
  • [21] Feature Selection from Barkhausen Noise Data Using Genetic Algorithms with Cross-Validation
    Sorsa, Aki
    Leiviska, Kauko
    ADAPTIVE AND NATURAL COMPUTING ALGORITHMS, 2009, 5495 : 213 - 222
  • [22] Measuring the bias of incorrect application of feature selection when using cross-validation in radiomics
    Aydin Demircioğlu
    Insights into Imaging, 12
  • [23] Dynamic weighting ensemble classifiers based on cross-validation
    Zhu Yu-Quan
    Ou Ji-Shun
    Chen Geng
    Yu Hai-Ping
    NEURAL COMPUTING & APPLICATIONS, 2011, 20 (03): : 309 - 317
  • [24] Dynamic weighting ensemble classifiers based on cross-validation
    Zhu Yu-Quan
    Ou Ji-Shun
    Chen Geng
    Yu Hai-Ping
    Neural Computing and Applications, 2011, 20 : 309 - 317
  • [25] Robust linear model selection by cross-validation
    Ronchetti, E
    Field, C
    Blanchard, W
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1997, 92 (439) : 1017 - 1023
  • [26] A survey of cross-validation procedures for model selection
    Arlot, Sylvain
    Celisse, Alain
    STATISTICS SURVEYS, 2010, 4 : 40 - 79
  • [27] MODEL-STRUCTURE SELECTION BY CROSS-VALIDATION
    STOICA, P
    EYKHOFF, P
    JANSSEN, P
    SODERSTROM, T
    INTERNATIONAL JOURNAL OF CONTROL, 1986, 43 (06) : 1841 - 1878
  • [29] Cross-validation for selecting a model selection procedure
    Zhang, Yongli
    Yang, Yuhong
    JOURNAL OF ECONOMETRICS, 2015, 187 (01) : 95 - 112
  • [30] MODEL SELECTION VIA MULTIFOLD CROSS-VALIDATION
    ZHANG, P
    ANNALS OF STATISTICS, 1993, 21 (01): : 299 - 313