Discriminant Mutual Information for Text Feature Selection

被引:1
作者
Wang, Jiaqi [1 ]
Zhang, Li [1 ,2 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou, Peoples R China
[2] Soochow Univ, Prov Key Lab Comp Informat Proc Technol, Suzhou, Peoples R China
来源
DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II | 2021年 / 12682卷
关键词
Text classification; Feature selection; Mutual information; Discriminant information; Redundant features; MACHINE;
D O I
10.1007/978-3-030-73197-7_9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In text classification tasks, the high dimensionality of data would result in a high computational complexity and decrease the classification accuracy because of high correlation between features; so, it is necessary to execute feature selection. In this paper, we propose a Discriminant Mutual Information (DMI) criterion to select features for text classification tasks. DMI measures the discriminant ability of features from two aspects. One is the mutual information between features and the label information. The other is the discriminant correlation degree between a feature and a target feature subset based on the label information, which could be used for judging whether a feature is redundant in the target feature subset. Thus, DMI is a de-redundancy text feature selection method considering discriminant information. In order to prove the superiority of DMI, we compare it with the state-of-the-art filter methods for text feature selection and conduct experiments on two datasets: Reuters-21578 and WebKB. K-Nearest Neighbor (KNN) and Support Vector Machine (SVM) are taken as the subsequent classifiers. Experimental results shows that the proposed DMI has significantly improved the classification accuracy and F1-score of both Reuters-21578 and WebKB.
引用
收藏
页码:136 / 151
页数:16
相关论文
共 23 条
[1]  
[Anonymous], 2017, arXiv
[2]  
[Anonymous], 2004, P 21 INT C MACH LEAR, DOI [10.1145/1015330.101535, DOI 10.1145/1015330.1015356]
[3]  
Cardoso-Cachopo A., 2007, PdD Thesis
[4]  
CHURCH KW, 1990, 27TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, P76
[5]  
Clark S., 2015, HDB CONT SEMANTIC TH, P493, DOI DOI 10.1002/9781118882139.CH16
[6]  
Cunningham, 2007, K NEAREST NEIGHBOUR, V34, P1
[7]   A probabilistic model derived term weighting scheme for text classification [J].
Feng, Guozhong ;
Li, Shaoting ;
Sun, Tieli ;
Zhang, Bangzuo .
PATTERN RECOGNITION LETTERS, 2018, 110 :23-29
[8]   MIFS-ND: A mutual information-based feature selection method [J].
Hoque, N. ;
Bhattacharyya, D. K. ;
Kalita, J. K. .
EXPERT SYSTEMS WITH APPLICATIONS, 2014, 41 (14) :6371-6385
[9]   Wrappers for feature subset selection [J].
Kohavi, R ;
John, GH .
ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) :273-324
[10]   A novel multivariate filter method for feature selection in text classification problems [J].
Labani, Mahdieh ;
Moradi, Parham ;
Ahmadizar, Fardin ;
Jalili, Mahdi .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2018, 70 :25-37