Feature selection for software effort estimation with localized neighborhood mutual information

被引:11
作者
Liu, Qin [1 ]
Xiao, Jiakai [2 ]
Zhu, Hongming [1 ]
机构
[1] Tongji Univ, Sch Software Engn, Shanghai 201804, Peoples R China
[2] Tongji Univ, Dept Comp Sci & Technol, Shanghai 201804, Peoples R China
来源
CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS | 2019年 / 22卷 / Suppl 3期
关键词
Feature selection; Case based reasoning; Neighborhood mutual information; Software effort estimation;
D O I
10.1007/s10586-018-1884-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Feature selection is usually employed before applying case based reasoning (CBR) for Software Effort Estimation (SEE). Unfortunately, most feature selection methods treat CBR as a black box method so there is no guarantee on the appropriateness of CBR on selected feature subset. The key to solve the problem is to measure the appropriateness of CBR assumption for a given feature set. In this paper, a measure called localized neighborhood mutual information (LNI) is proposed for this purpose and a greedy method called LNI based feature selection (LFS) is designed for feature selection. Experiment with leave-one-out cross validation (LOOCV) on 6 benchmark datasets demonstrates that: (1) CBR makes effective estimation with the LFS selected subset compared with a randomized baseline method. Compared with three representative feature selection methods, (2) LFS achieves optimal MAR value on 3 out of 6 datasets with a 14% average improvement and (3) LFS achieves optimal MMRE on 5 out of 6 datasets with a 24% average improvement.
引用
收藏
页码:S6953 / S6961
页数:9
相关论文
共 50 条
[21]   Is mutual information adequate for feature selection in regression? [J].
Frenay, Benoit ;
Doquire, Gauthier ;
Verleysen, Michel .
NEURAL NETWORKS, 2013, 48 :1-7
[22]   Discriminant Mutual Information for Text Feature Selection [J].
Wang, Jiaqi ;
Zhang, Li .
DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II, 2021, 12682 :136-151
[23]   FEATURE SELECTION WITH WEIGHTED CONDITIONAL MUTUAL INFORMATION [J].
Celik, Ceyhun ;
Bilge, Hasan Sakir .
JOURNAL OF THE FACULTY OF ENGINEERING AND ARCHITECTURE OF GAZI UNIVERSITY, 2015, 30 (04) :585-596
[24]   Feature Selection with Mutual Information for Regression Problems [J].
Sulaiman, Muhammad Aliyu ;
Labadin, Jane .
2015 9TH INTERNATIONAL CONFERENCE ON IT IN ASIA (CITA), 2015,
[25]   Gait Feature Subset Selection by Mutual Information [J].
Guo, Baofeng ;
Nixon, Mark S. .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART A-SYSTEMS AND HUMANS, 2009, 39 (01) :36-46
[26]   Genetic algorithm for feature selection with mutual information [J].
Ge, Hong ;
Hu, Tianliang .
2014 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2014), VOL 1, 2014, :116-119
[27]   Feature Selection by Maximizing Part Mutual Information [J].
Gao, Wanfu ;
Hu, Liang ;
Zhang, Ping .
2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND MACHINE LEARNING (SPML 2018), 2018, :120-127
[28]   AFIFC: Adaptive fuzzy neighborhood mutual information-based feature selection via label correlation [J].
Sun, Lin ;
Xu, Feng ;
Ding, Weiping ;
Xu, Jiucheng .
PATTERN RECOGNITION, 2025, 164
[29]   Feature Selection with Conditional Mutual Information Considering Feature Interaction [J].
Liang, Jun ;
Hou, Liang ;
Luan, Zhenhua ;
Huang, Weiping .
SYMMETRY-BASEL, 2019, 11 (07)
[30]   Feature Selection Technique for Effective Software Effort Estimation Using Multi-Layer Perceptrons [J].
Goyal, Somya ;
Bhatia, Pradeep K. .
PROCEEDINGS OF ICETIT 2019: EMERGING TRENDS IN INFORMATION TECHNOLOGY, 2020, 605 :181-192