Regularized Multivariate Analysis Framework for Interpretable High-Dimensional Variable Selection

被引:10
|
作者
Munoz-Romero, Sergio [1 ]
Gomez-Verdejo, Vanessa [2 ]
Arenas-Garcia, Jernimo [2 ]
机构
[1] Univ Rey Juan Carlos, Dept Signal Proc & Commun, Madrid, Spain
[2] Univ Carlos III Madrid, Dept Signal Proc & Commun, E-28903 Getafe, Spain
关键词
SPARSE; REGRESSION;
D O I
10.1109/MCI.2016.2601701
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multivariate Analysis (MVA) comprises a family of well-known methods for feature extraction which exploit correlations among input variables representing the data. One important property that is enjoyed by most such methods is uncorrelation among the extracted features. Recently, regularized versions of MVA methods have appeared in the literature, mainly with the goal to gain interpretability of the solution. In these cases, the solutions can no longer be obtained in a closed manner, and more complex optimization methods that rely on the iteration of two steps are frequently used. This paper recurs to an alternative approach to solve efficiently this iterative problem. The main novelty of this approach lies in preserving several properties of the original methods, most notably the uncorrelation of the extracted features. Under this framework, we propose a novel method that takes advantage of the,2,1 norm to perform variable selection during the feature extraction process. Experimental results over different problems corroborate the advantages of the proposed formulation in comparison to state of the art formulations.
引用
收藏
页码:24 / 35
页数:12
相关论文
共 50 条
  • [41] The use of random-effect models for high-dimensional variable selection problems
    Kwon, Sunghoon
    Oh, Seungyoung
    Lee, Youngjo
    COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2016, 103 : 401 - 412
  • [42] Bayesian variable selection in multinomial probit model for classifying high-dimensional data
    Yang, Aijun
    Li, Yunxian
    Tang, Niansheng
    Lin, Jinguan
    COMPUTATIONAL STATISTICS, 2015, 30 (02) : 399 - 418
  • [43] Posterior model consistency in high-dimensional Bayesian variable selection with arbitrary priors
    Hua, Min
    Goh, Gyuhyeong
    STATISTICS & PROBABILITY LETTERS, 2025, 223
  • [44] Fast and Scalable Spike and Slab Variable Selection in High-Dimensional Gaussian Processes
    Dance, Hugh
    Paige, Brooks
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [45] Empirical Study on High-Dimensional Variable Selection and Prediction Under Competing Risks
    Hou, Jiayi
    Xu, Ronghui
    NEW FRONTIERS OF BIOSTATISTICS AND BIOINFORMATICS, 2018, : 421 - 440
  • [46] Transformed low-rank ANOVA models for high-dimensional variable selection
    Jung, Yoonsuh
    Zhang, Hong
    Hu, Jianhua
    STATISTICAL METHODS IN MEDICAL RESEARCH, 2019, 28 (04) : 1230 - 1246
  • [47] Post selection shrinkage estimation for high-dimensional data analysis
    Gao, Xiaoli
    Ahmed, S. E.
    Feng, Yang
    APPLIED STOCHASTIC MODELS IN BUSINESS AND INDUSTRY, 2017, 33 (02) : 97 - 120
  • [48] Doubly regularized estimation and selection in linear mixed-effects models for high-dimensional longitudinal data
    Li, Yun
    Wang, Sijian
    Song, Peter X-K
    Wang, Naisyin
    Zhou, Ling
    Zhu, Ji
    STATISTICS AND ITS INTERFACE, 2018, 11 (04) : 721 - 737
  • [49] Variable Selection Using Nonlocal Priors in High-Dimensional Generalized Linear Models With Application to fMRI Data Analysis
    Cao, Xuan
    Lee, Kyoungjae
    ENTROPY, 2020, 22 (08)
  • [50] Variable selection for high-dimensional generalized linear model with block-missing data
    He, Yifan
    Feng, Yang
    Song, Xinyuan
    SCANDINAVIAN JOURNAL OF STATISTICS, 2023, 50 (03) : 1279 - 1297