Interpretation of linear classifiers by means of feature relevance bounds

被引:7
作者
Goepfert, Christina [1 ]
Pfannschmidt, Lukas [1 ]
Goepfert, Jan Philip [1 ]
Hammer, Barbara [1 ]
机构
[1] Cognit Interact Technol, Inspirat 1, D-33619 Bielefeld, Germany
关键词
Feature relevance; Feature selection; Interpretability; All-relevant; Linear classification; FEATURE-SELECTION;
D O I
10.1016/j.neucom.2017.11.074
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Research on feature relevance and feature selection problems goes back several decades, but the importance of these areas continues to grow as more and more data becomes available, and machine learning methods are used to gain insight and interpret, rather than solely to solve classification or regression problems. Despite the fact that feature relevance is often discussed, it is frequently poorly defined, and the feature selection problems studied are subtly different. Furthermore, the problem of finding all features relevant for a classification problem has only recently started to gain traction, despite its importance for interpretability and integrating expert knowledge. In this paper, we attempt to unify commonly used concepts and to give an overview of the main questions and results. We formalize two interpretations of the all-relevant problem and propose a polynomial method to approximate one of them for the important hypothesis class of linear classifiers, which also enables a distinction between strongly and weakly relevant features. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:69 / 79
页数:11
相关论文
共 50 条
  • [31] Finding optimal classifiers for small feature sets in genomics and proteomics
    Stiglic, Gregor
    Rodriguez, Juan J.
    Kokol, Peter
    NEUROCOMPUTING, 2010, 73 (13-15) : 2346 - 2352
  • [32] A straightforward feature selection method based on mean ratio for classifiers
    Tavares, Emmanuel
    Silva, Alisson Marques
    Moita, Gray Farias
    Nogueira Cardoso, Rodrigo Tomas
    INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2021, 15 (03): : 421 - 432
  • [33] Bias and stability of single variable classifiers for feature ranking and selection
    Fakhraei, Shobeir
    Soltanian-Zadeh, Hamid
    Fotouhi, Farshad
    EXPERT SYSTEMS WITH APPLICATIONS, 2014, 41 (15) : 6945 - 6958
  • [34] Feature selection, Optimization and Performance Analysis of Classifiers for Biological Images
    Siji, K. K.
    Mathew, Binitha Sara
    Chandran, Rakhi
    Shajeemohan, B. S.
    Shanthini, K. S.
    2014 NATIONAL CONFERENCE ON COMMUNICATION, SIGNAL PROCESSING AND NETWORKING (NCCSN), 2014,
  • [35] A Study on Mutual Information-Based Feature Selection in Classifiers
    Arundhathi, B.
    Athira, A.
    Rajan, Ranjidha
    ARTIFICIAL INTELLIGENCE AND EVOLUTIONARY COMPUTATIONS IN ENGINEERING SYSTEMS, ICAIECES 2016, 2017, 517 : 479 - 486
  • [36] A New Feature Selection Approach to Naive Bayes Text Classifiers
    Zhang, Lungan
    Jiang, Liangxiao
    Li, Chaoqun
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2016, 30 (02)
  • [37] Feature Relevance in Ward's Hierarchical Clustering Using the L p Norm
    de Amorim, Renato Cordeiro
    JOURNAL OF CLASSIFICATION, 2015, 32 (01) : 46 - 62
  • [38] Relevance-based label distribution feature selection via convex optimization
    Qian, Wenbin
    Ye, Qianzhi
    Li, Yihui
    Huang, Jintao
    Dai, Shiming
    INFORMATION SCIENCES, 2022, 607 : 322 - 345
  • [39] Multi-Label Feature Selection Combining Three Types of Conditional Relevance
    Gao, Lingbo
    Wang, Yiqiang
    Li, Yonghao
    Zhang, Ping
    Hu, Liang
    ENTROPY, 2021, 23 (12)
  • [40] Assessing feature relevance in NPLS models by VIP
    Favilla, Stefania
    Durante, Caterina
    Vigni, Mario Li
    Cocchi, Marina
    CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 2013, 129 : 76 - 86