Sparse optimization in feature selection: application in neuroimaging

被引:0
|
作者
K. Kampa
S. Mehta
C. A. Chou
W. A. Chaovalitwongse
T. J. Grabowski
机构
[1] University of Washington,Department of Industrial and Systems Engineering
[2] University of Washington Medical Center,Integrated Brain Imaging Center
[3] University of Washington,Department of Radiology
[4] University of Washington,Department of Psychology
[5] Binghamton University,Department of Systems Science and Industrial Engineering
[6] State University of New York,Department of Radiology
[7] University of Washington,Department of Neurology
[8] University of Washington,undefined
来源
Journal of Global Optimization | 2014年 / 59卷
关键词
Sparse optimization; Feature selection; Machine learning; fMRI; Cognitive neuroscience; Regularization ; Pattern classification;
D O I
暂无
中图分类号
学科分类号
摘要
Feature selection plays an important role in the successful application of machine learning techniques to large real-world datasets. Avoiding model overfitting, especially when the number of features far exceeds the number of observations, requires selecting informative features and/or eliminating irrelevant ones. Searching for an optimal subset of features can be computationally expensive. Functional magnetic resonance imaging (fMRI) produces datasets with such characteristics creating challenges for applying machine learning techniques to classify cognitive states based on fMRI data. In this study, we present an embedded feature selection framework that integrates sparse optimization for regularization (or sparse regularization) and classification. This optimization approach attempts to maximize training accuracy while simultaneously enforcing sparsity by penalizing the objective function for the coefficients of the features. This process allows many coefficients to become zero, which effectively eliminates their corresponding features from the classification model. To demonstrate the utility of the approach, we apply our framework to three different real-world fMRI datasets. The results show that regularized classifiers yield better classification accuracy, especially when the number of initial features is large. The results further show that sparse regularization is key to achieving scientifically-relevant generalizability and functional localization of classifier features. The approach is thus highly suited for analysis of fMRI data.
引用
收藏
页码:439 / 457
页数:18
相关论文
共 50 条
  • [1] Sparse optimization in feature selection: application in neuroimaging
    Kampa, K.
    Mehta, S.
    Chou, C. A.
    Chaovalitwongse, W. A.
    Grabowski, T. J.
    JOURNAL OF GLOBAL OPTIMIZATION, 2014, 59 (2-3) : 439 - 457
  • [2] Discriminative sparse subspace learning and its application to unsupervised feature selection
    Zhou, Nan
    Cheng, Hong
    Pedrycz, Witold
    Zhang, Yong
    Liu, Huaping
    ISA TRANSACTIONS, 2016, 61 : 104 - 118
  • [3] Sparse optimization via vector k-norm and DC programming with an application to feature selection for support vector machines
    Gaudioso, Manlio
    Giallombardo, Giovanni
    Miglionico, Giovanna
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2023, 86 (02) : 745 - 766
  • [4] Sparse Support Vector Machine with Lp Penalty for Feature Selection
    Lan Yao
    Feng Zeng
    Dong-Hui Li
    Zhi-Gang Chen
    Journal of Computer Science and Technology, 2017, 32 : 68 - 77
  • [5] Sparse Graph Embedding Unsupervised Feature Selection
    Wang, Shiping
    Zhu, William
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2018, 48 (03): : 329 - 341
  • [6] UDSFS: Unsupervised deep sparse feature selection
    Cong, Yang
    Wang, Shuai
    Fan, Baojie
    Yang, Yunsheng
    Yu, Haibin
    NEUROCOMPUTING, 2016, 196 : 150 - 158
  • [7] Binary feature mask optimization for feature selection
    Lorasdagi, Mehmet E.
    Turali, Mehmet Y.
    Kozat, Suleyman S.
    Neural Computing and Applications, 2025, 37 (06) : 5155 - 5167
  • [8] An application of MOGW optimization for feature selection in text classification
    Razieh Asgarnezhad
    S. Amirhassan Monadjemi
    Mohammadreza Soltanaghaei
    The Journal of Supercomputing, 2021, 77 : 5806 - 5839
  • [9] Application of global optimization methods to model and feature selection
    Boubezoul, Abderrahmane
    Paris, Sebastien
    PATTERN RECOGNITION, 2012, 45 (10) : 3676 - 3686
  • [10] An application of MOGW optimization for feature selection in text classification
    Asgarnezhad, Razieh
    Monadjemi, S. Amirhassan
    Soltanaghaei, Mohammadreza
    JOURNAL OF SUPERCOMPUTING, 2021, 77 (06) : 5806 - 5839