Feature selection with scalable variational gaussian process via sensitivity analysis based on L2 divergence

被引:2
|
作者
Jeon, Younghwan [1 ]
Hwang, Ganguk [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Dept Math Sci, Daejeon 305701, South Korea
基金
新加坡国家研究基金会;
关键词
Gaussian processes; Scalable variational gaussian process; Feature selection; L2; divergence; VARIABLE SELECTION; MODELS;
D O I
10.1016/j.neucom.2022.11.013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection is one of the most important issues in supervised learning and there are a lot of different feature selection approaches in the literature. Among them one recent approach is to use Gaussian pro-cess (GP) because it can capture well the hidden relevance between the features of the input and the out-put. However, the existing feature selection approaches with GP suffer from the scalability problem due to high computational cost of inference with GP. Moreover, they use the Kullback-Leibler (KL) divergence in the sensitivity analysis for feature selection, but we show in this paper that the KL divergence under-estimates the relevance of important features in some cases of classification. To remedy such drawbacks of the existing GP based approaches, we propose a new feature selection method with scalable variational Gaussian process (SVGP) and L2 divergence. With the help of SVGP the proposed method exploits given large data sets well for feature selection through so-called inducing points while avoiding the scalability problem. Moreover, we provide theoretical analysis to motivate the choice of L2 divergence for feature selection in both classification and regression. To validate the perfor-mance of the proposed method, we compare it with other existing methods through experiments with synthetic and real data sets. (c) 2022 Elsevier B.V. All rights reserved.
引用
收藏
页码:577 / 592
页数:16
相关论文
共 50 条
  • [31] Honest variable selection in linear and logistic regression models via l1 and l1 + l2 penalization
    Bunea, Florentina
    ELECTRONIC JOURNAL OF STATISTICS, 2008, 2 : 1153 - 1194
  • [32] Unsupervised maximum margin feature selection via L 2,1-norm minimization
    Yang, Shizhun
    Hou, Chenping
    Nie, Feiping
    Wu, Yi
    NEURAL COMPUTING & APPLICATIONS, 2012, 21 (07) : 1791 - 1799
  • [33] Face recognition based on selection approach via Canonical Correlation Analysis feature fusion
    Huy Nguyen-Quoc
    Vinh Truong Hoang
    2020 ZOOMING INNOVATION IN CONSUMER TECHNOLOGIES CONFERENCE (ZINC), 2020, : 54 - 57
  • [34] Genetic Algorithm Based Feature Selection on Diagnosis of Parkinson Disease via Vocal Analysis
    Gumuscu, Abdulkadir
    Karadag, Kerim
    Tenekeci, Mehmet Emin
    Aydilek, Ibrahim Berkan
    2017 25TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2017,
  • [35] Enhancing Phishing Website Detection via Feature Selection in URL-Based Analysis
    Qasim M.A.
    Flayh N.A.
    Informatica (Slovenia), 2023, 47 (09): : 145 - 155
  • [36] Low-resolution face recognition and feature selection based on multidimensional scaling joint L2,1-norm regularisation
    Li, Fei
    Jiang, Mingyan
    IET BIOMETRICS, 2019, 8 (03) : 198 - 205
  • [37] Machine learning framework for breast cancer detection with feature selection with L2 ridge regularization: insights from multiple datasets
    Kandhasamy, Premalatha
    Devi, Duraisamy Prabha
    Kandhasamy, Sivakumar
    JOURNAL OF TRANSLATIONAL GENETICS AND GENOMICS, 2025, 9 : 11 - 34
  • [38] NCMS: Towards accurate anchor free object detection through l2 norm calibration and multi-feature selection
    Chen, Fangyi
    Zhu, Chenchen
    Shen, Zhiqiang
    Zhang, Han
    Savvides, Marios
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2020, 200
  • [39] Prediction using step-wise L1, L2 regularization and feature selection for small data sets with large number of features
    Ozgur Demir-Kavuk
    Mayumi Kamada
    Tatsuya Akutsu
    Ernst-Walter Knapp
    BMC Bioinformatics, 12
  • [40] A novel feature selection method based on global sensitivity analysis with application in machine learning-based prediction model
    Zhang, Pin
    APPLIED SOFT COMPUTING, 2019, 85