A New Approach for Automated Feature Selection

被引:0
|
作者
Gocht, Andreas [1 ]
Lehmann, Christoph [1 ]
Schoene, Robert [1 ]
机构
[1] Tech Univ Dresden, Ctr Informat Serv & High Performance Comp ZIH, D-01062 Dresden, Germany
来源
2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA) | 2018年
基金
欧盟地平线“2020”;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection or variable selection is an important step in different machine learning tasks. In a traditional approach, users specify the amount of features, which shall be selected. Afterwards, algorithm select features by using scores like the Joint Mutual Information (JMI). If users do not know the exact amount of features to select, they need to evaluate the full learning chain for different feature counts in order to determine, which amount leads to the lowest training error. To overcome this drawback, we extend the JMI score and mitigate the flaw by introducing a stopping criterion to the selection algorithm that can be specified depending on the learning task. With this, we enable developers to carry out the feature selection task before the actual learning is done. We call our new score Historical Joint Mutual Information (HJMI). Additionally, we compare our new algorithm, using the novel HJMI score, against traditional algorithms, which use the JMI score. With this, we demonstrate that the HJMI-based algorithm is able to automatically select a reasonable amount of features: Our approach delivers results as good as traditional approaches and sometimes even outperforms them, as it is not limited to a certain step size for feature evaluation.
引用
收藏
页码:4915 / 4920
页数:6
相关论文
共 50 条
  • [1] An automated parameter selection approach for simultaneous clustering and feature selection
    Kumar, Vijay
    Chhabra, Jitender K.
    Kumar, Dinesh
    JOURNAL OF ENGINEERING RESEARCH, 2016, 4 (02): : 65 - 85
  • [2] A new approach to feature selection
    Scherf, M
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 1997, 1211 : 181 - 184
  • [3] A Feature-Free Approach to Automated Algorithm Selection
    Alissa, Mohamad
    Sim, Kevin
    Hart, Emma
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 9 - 10
  • [4] A new graph feature selection approach
    Akhiat, Yassine
    Asnaoui, Youssef
    Chahhou, Mohamed
    Zinedine, Ahmed
    2020 6TH IEEE CONGRESS ON INFORMATION SCIENCE AND TECHNOLOGY (IEEE CIST'20), 2020, : 156 - 161
  • [5] A new approach to feature subset selection
    Liu, DZ
    Feng, ZJ
    Wang, XZ
    PROCEEDINGS OF THE 2004 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2004, : 1822 - 1825
  • [6] A new approach to feature selection in text classification
    Wang, Y
    Wang, XJ
    PROCEEDINGS OF 2005 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-9, 2005, : 3814 - 3819
  • [7] AIFSA: A New Approach for Feature Selection and Weighting
    Fouad, Walid
    Badr, Amr
    Farag, Ibrahim
    INFORMATICS ENGINEERING AND INFORMATION SCIENCE, PT II, 2011, 252 : 596 - 609
  • [8] A New Approach of Feature Selection for Text Categorization
    CUI Zifeng~1
    2. Department of Computer Science and Engineering
    WuhanUniversityJournalofNaturalSciences, 2006, (05) : 1335 - 1339
  • [9] A new approach to feature selection for text categorization
    Li, SS
    Zong, CQ
    PROCEEDINGS OF THE 2005 IEEE INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND KNOWLEDGE ENGINEERING (IEEE NLP-KE'05), 2005, : 626 - 630
  • [10] Automated feature selection in neuroevolution
    Tan, Maxine
    Hartley, Michael
    Bister, Michel
    Deklerck, Rudi
    EVOLUTIONARY INTELLIGENCE, 2009, 1 (04) : 271 - 292