Class-specific feature selection using fuzzy information-theoretic metrics

被引:7
作者
Ma, Xi-Ao [1 ,2 ]
Xu, Hao [1 ]
Liu, Yi [3 ]
Zhang, Justin Zuopeng [4 ]
机构
[1] Zhejiang Gongshang Univ, Sch Comp Sci & Technol, Hangzhou 310018, Zhejiang, Peoples R China
[2] Zhejiang Gongshang Univ, Collaborat Innovat Ctr Computat Social Sci, Hangzhou 310018, Zhejiang, Peoples R China
[3] Zhejiang Sci Tech Univ, Sch Econ & Management, Hangzhou 310018, Zhejiang, Peoples R China
[4] Univ North Florida, Coggin Coll Business, Jacksonville, FL 32224 USA
关键词
Feature selection; Fuzzy information-theoretic metric; Feature relevance; Feature redundancy; Class-specific feature selection; MUTUAL INFORMATION; MAX-RELEVANCE; REDUNDANCY; REDUCTION; ALGORITHM;
D O I
10.1016/j.engappai.2024.109035
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Fuzzy information-theoretic metrics have been demonstrated to be effective in evaluating feature relevance and redundancy in both categorical and numerical feature selection tasks. Most existing feature selection methods based on fuzzy information-theoretic metrics treat all classes as a single entity, resulting in the selection of a consistent feature subset for all classes. This approach overlooks the fact that diverse classes may exhibit distinct discriminative characteristics, which requires the selection of diverse feature subsets for each class. Consequently, these methods lack the capability to handle this variability in feature selection required for optimal performance. To address this limitation, this paper proposes a class-specific feature selection method based on fuzzy information-theoretic metrics. To be more specific, we introduce several class-specific fuzzy information-theoretic metrics. Building upon these metrics, we formulate a class-specific feature selection algorithm, which is referred to as class-specific fuzzy information-theoretic feature selection. This algorithm enables the selection of highly relevant feature subsets tailored to each individual class. Furthermore, we present a class-specific ensemble classification framework that integrates the classification results obtained from the feature subsets generated by our method. Finally, we conduct extensive tests on 21 publicly available datasets using four popular classifiers to compare the performance of our method with eight up-to-date classification-specific methods. Test results demonstrate that our method outperforms the eight compared classification-specific methods in terms of performance.
引用
收藏
页数:16
相关论文
共 62 条
[1]  
Bagenstoss PM, 1999, IEEE T SIGNAL PROCES, V47, P3428, DOI 10.1109/78.806092
[2]   Feature selection using Joint Mutual Information Maximisation [J].
Bennasar, Mohamed ;
Hicks, Yulia ;
Setchi, Rossitza .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (22) :8520-8532
[3]   Ensembles for feature selection: A review and future trends [J].
Bolon-Canedo, Veronica ;
Alonso-Betanzos, Amparo .
INFORMATION FUSION, 2019, 52 :1-12
[4]   Benchmark for filter methods for feature selection in high-dimensional classification data [J].
Bommert, Andrea ;
Sun, Xudong ;
Bischl, Bernd ;
Rahnenfuehrer, Joerg ;
Lang, Michel .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2020, 143
[5]   Improvement of K-means Cluster Quality by Post Processing Resulted Clusters [J].
Borlea, Ioan-Daniel ;
Precup, Radu-Emil ;
Borlea, Alexandra-Bianca .
8TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND QUANTITATIVE MANAGEMENT (ITQM 2020 & 2021): DEVELOPING GLOBAL DIGITAL ECONOMY AFTER COVID-19, 2022, 199 :63-70
[6]   Automatic ECG arrhythmias classification scheme based on the conjoint use of the multi-layer perceptron neural network and a new improved metaheuristic approach [J].
Bouaziz, Fatiha ;
Oulhadj, Hamouche ;
Boutana, Daoud ;
Siarry, Patrick .
IET SIGNAL PROCESSING, 2019, 13 (08) :726-735
[7]   A PSO-aided neuro-fuzzy classifier employing linguistic hedge concepts [J].
Chatterjee, Amitava ;
Siarry, Patrick .
EXPERT SYSTEMS WITH APPLICATIONS, 2007, 33 (04) :1097-1109
[8]   Feature selection via normative fuzzy information weight with application into tumor classification [J].
Dai, Jianhua ;
Chen, Jiaolong .
APPLIED SOFT COMPUTING, 2020, 92
[9]   Attribute selection based on information gain ratio in fuzzy rough set theory with application to tumor classification [J].
Dai, Jianhua ;
Xu, Qing .
APPLIED SOFT COMPUTING, 2013, 13 (01) :211-221
[10]   Pointwise mutual information sparsely embedded feature selection [J].
Deng, Tingquan ;
Huang, Yang ;
Yang, Ge ;
Wang, Changzhong .
INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2022, 151 :251-270