Optimized Feature Selection Using Modified Social Group Optimization

被引:0
作者
Meesala, Y.V. Nagesh [1 ]
Parida, Ajaya Kumar [2 ]
Naik, Anima [3 ]
机构
[1] School of Computer Engineering, KIIT Deemed to be University, Department of CSE, Raghu Engineering College
[2] School of Computer Engineering, KIIT Deemed to be University
[3] Department of CSE, Raghu Engineering College, Andhra Pradesh, Visakhapatnam
来源
Informatica (Slovenia) | 2024年 / 48卷 / 11期
关键词
CA; datasets; FS; optimization algorithm; SGO;
D O I
10.31449/inf.v48i11.6160
中图分类号
学科分类号
摘要
This paper introduces binary variants of the Modified Social Group Optimization (MSGO) algorithm designed specifically for optimal feature subset selection in a wrapper-mode classification setting. While the original SGO was proposed in 2016 and modified in 2020 to enhance its performance, it was not previously applied to feature selection problems. MSGO represents an advancement over SGO, adept at efficiently exploring the feature space to identify optimal or near-optimal feature subsets by minimizing a specified fitness function. The two newly proposed binary variants of MSGO are employed to identify the optimal feature combinations that maximize classification accuracy while minimizing the number of selected features. In these variants, the native MSGO is utilized while its continuous steps are bounded in a threshold using a suitable threshold function after squashing them. These binary algorithms are compared against six latest high-performing optimization approaches and six state-of-the-art optimization algorithms to assess their performance. Various evaluation metrics are utilized across twenty-three datasets sourced from the UCI data repository to accurately judge and compare the efficacy of these algorithms. The experimental results confirm the efficiency of the proposed approaches in improving the classification accuracy compared to other wrapper-based algorithms, which proves the ability of the MSGO algorithm to search the feature space and select the most informative attributes for classification tasks. © 2024 Slovene Society Informatika. All rights reserved.
引用
收藏
页码:195 / 220
页数:25
相关论文
共 60 条
[1]  
Bennasar M., Hicks Y., Setchi R., Feature selection using Joint Mutual Information Maximisation, Expert Systems with Applications, 42, 22, pp. 8520-8532, (2015)
[2]  
Zawbaa H. M., Emary E., Grosan C., Feature selection via chaotic antlion optimization, PloS One, 11, 3, (2016)
[3]  
Liu H., Yu L., Toward integrating feature selection algorithms for classification and clustering, IEEE Transactions on Knowledge and Data Engineering, 17, 4, pp. 491-502, (2005)
[4]  
Yang Y., Pedersen J.O., A comparative study on feature selection in text categorization, Icml, 97, pp. 412-420, (1997)
[5]  
Naik A., Satapathy S. C., Efficient clustering of dataset based on particle swarm optimization, International Journal of Computer Science Engineering and Information Technology Research (IJCSEITR), 3, 1, pp. 39-48, (2013)
[6]  
Naik Anima, Satapathy S. C., Efficient clustering of dataset based on differential evolution, Advances in Intelligent Systems and Computing. Proceedings of the International Conference on Frontiers of Intelligent Computing: Theory and Applications (FICTA) 2013, pp. 217-227, (2014)
[7]  
Han J., Kamber M., Pei J., Data Preprocessing, Data Mining, pp. 83-124, (2012)
[8]  
Peng H., Long F., Ding C., Feature selection based on mutual information: criteria of max-dependency, max-relevance, and min-redundancy, IEEE Transactions on Pattern Analysis and Machine Intelligence, 27, 8, pp. 1226-1238, (2005)
[9]  
Kira K., Rendell L. A., A practical approach to feature selection, Machine Learning Proceedings 1992, pp. 249-256, (1992)
[10]  
Zheng Z., Wu X., Srihari R., Feature selection for text categorization on imbalanced data, SIGKDD Explorations: Newsletter of the Special Interest Group (SIG) on Knowledge Discovery & Data Mining, 6, 1, pp. 80-89, (2004)