Sparse optimization in feature selection: application in neuroimaging

被引:0
|
作者
K. Kampa
S. Mehta
C. A. Chou
W. A. Chaovalitwongse
T. J. Grabowski
机构
[1] University of Washington,Department of Industrial and Systems Engineering
[2] University of Washington Medical Center,Integrated Brain Imaging Center
[3] University of Washington,Department of Radiology
[4] University of Washington,Department of Psychology
[5] Binghamton University,Department of Systems Science and Industrial Engineering
[6] State University of New York,Department of Radiology
[7] University of Washington,Department of Neurology
[8] University of Washington,undefined
来源
Journal of Global Optimization | 2014年 / 59卷
关键词
Sparse optimization; Feature selection; Machine learning; fMRI; Cognitive neuroscience; Regularization ; Pattern classification;
D O I
暂无
中图分类号
学科分类号
摘要
Feature selection plays an important role in the successful application of machine learning techniques to large real-world datasets. Avoiding model overfitting, especially when the number of features far exceeds the number of observations, requires selecting informative features and/or eliminating irrelevant ones. Searching for an optimal subset of features can be computationally expensive. Functional magnetic resonance imaging (fMRI) produces datasets with such characteristics creating challenges for applying machine learning techniques to classify cognitive states based on fMRI data. In this study, we present an embedded feature selection framework that integrates sparse optimization for regularization (or sparse regularization) and classification. This optimization approach attempts to maximize training accuracy while simultaneously enforcing sparsity by penalizing the objective function for the coefficients of the features. This process allows many coefficients to become zero, which effectively eliminates their corresponding features from the classification model. To demonstrate the utility of the approach, we apply our framework to three different real-world fMRI datasets. The results show that regularized classifiers yield better classification accuracy, especially when the number of initial features is large. The results further show that sparse regularization is key to achieving scientifically-relevant generalizability and functional localization of classifier features. The approach is thus highly suited for analysis of fMRI data.
引用
收藏
页码:439 / 457
页数:18
相关论文
共 50 条
  • [31] Sparse Coding for Feature Selection on Genome-Wide Association Data
    Braenne, Ingrid
    Labusch, Kai
    Mamlouk, Amir Madany
    ARTIFICIAL NEURAL NETWORKS-ICANN 2010, PT I, 2010, 6352 : 337 - 346
  • [32] Sparse Support Vector Machine with L p Penalty for Feature Selection
    Yao, Lan
    Zeng, Feng
    Li, Dong-Hui
    Chen, Zhi-Gang
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2017, 32 (01) : 68 - 77
  • [33] Parallel binary arithmetic optimization algorithm and its application for feature selection
    Zhuang, Zhongjie
    Pan, Jeng-Shyang
    Li, Junbao
    Chu, Shu-Chuan
    KNOWLEDGE-BASED SYSTEMS, 2023, 275
  • [34] A Survey on Sparse Learning Models for Feature Selection
    Li, Xiaoping
    Wang, Yadi
    Ruiz, Ruben
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (03) : 1642 - 1660
  • [35] FEATURE SELECTION USING PARTICLE SWARM OPTIMIZATION WITH APPLICATION IN SPAM FILTERING
    Lai, Chih-Chin
    Wu, Chih-Hung
    Tsai, Ming-Chi
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2009, 5 (02): : 423 - 432
  • [36] Feature selection for portfolio optimization
    Thomas Trier Bjerring
    Omri Ross
    Alex Weissensteiner
    Annals of Operations Research, 2017, 256 : 21 - 40
  • [37] Feature selection for portfolio optimization
    Bjerring, Thomas Trier
    Ross, Omri
    Weissensteiner, Alex
    ANNALS OF OPERATIONS RESEARCH, 2017, 256 (01) : 21 - 40
  • [38] FEATURE SELECTION FOR APPLICATION RECOGNITION IN COMMUNICATION NETWORKS
    Zelina, Milan
    Oravec, Milos
    AD ALTA-JOURNAL OF INTERDISCIPLINARY RESEARCH, 2011, 1 (01): : 115 - 117
  • [39] Elite-driven grey wolf optimization for global optimization and its application to feature selection
    Zhang, Li
    Chen, Xiaobo
    SWARM AND EVOLUTIONARY COMPUTATION, 2025, 92
  • [40] A Multicriteria Optimization Approach for the Stock Market Feature Selection
    Radojicic, Dragana
    Radojicic, Nina
    Kredatus, Simeon
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2021, 18 (03) : 749 - 769