Improved Mutual Information Method For Text Feature Selection

被引:0
|
作者
Ding Xiaoming [1 ]
Tang Yan [1 ]
机构
[1] Southwest Univ, Coll Comp & Informat Sci, Chongqing 400715, Peoples R China
来源
PROCEEDINGS OF THE 2013 8TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE & EDUCATION (ICCSE 2013) | 2013年
关键词
text classification; feature selection; mutual information;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Reducing the dimensions of high-dimensional feature set is one of the difficulties of text categorization. Feature selection has been effectively applied in text classification, because of its low complexity of computing. Research works show that mutual information is a good feature selection method but doesn't consider the term frequency in each category of the corpus and the connections between terms. To remedying the defects of traditional mutual information method, this article improved measure of mutual information by introducing the feature frequency in class and the dispersion of feature in class, and built a experimental platform by constructing a Chinese text classification system, and did a multi-set of experiments base on this system. The results show that the new feature selection approach has a more excellent effect in text categorization.
引用
收藏
页码:163 / 166
页数:4
相关论文
共 50 条
  • [31] An improved feature selection method based on angle-guided multi-objective PSO and feature-label mutual information
    Fei Han
    Tianyi Wang
    Qinghua Ling
    Applied Intelligence, 2023, 53 : 3545 - 3562
  • [32] Is mutual information adequate for feature selection in regression?
    Frenay, Benoit
    Doquire, Gauthier
    Verleysen, Michel
    NEURAL NETWORKS, 2013, 48 : 1 - 7
  • [33] Genetic algorithm for feature selection with mutual information
    Ge, Hong
    Hu, Tianliang
    2014 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2014), VOL 1, 2014, : 116 - 119
  • [34] Feature Selection by Maximizing Part Mutual Information
    Gao, Wanfu
    Hu, Liang
    Zhang, Ping
    2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND MACHINE LEARNING (SPML 2018), 2018, : 120 - 127
  • [35] Mutual information for feature selection: estimation or counting?
    Nguyen H.B.
    Xue B.
    Andreae P.
    Evolutionary Intelligence, 2016, 9 (3) : 95 - 110
  • [36] Feature Selection with Mutual Information for Regression Problems
    Sulaiman, Muhammad Aliyu
    Labadin, Jane
    2015 9TH INTERNATIONAL CONFERENCE ON IT IN ASIA (CITA), 2015,
  • [37] Text Feature Selection Method in battlefield information service
    Wang Kai
    Liu Jingzhi
    Wang Kai
    Gan Zhichun
    Cai Yanjun
    2016 17TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED COMPUTING, APPLICATIONS AND TECHNOLOGIES (PDCAT), 2016, : 216 - 220
  • [38] Input Feature Selection Method Based on Feature Set Equivalence and Mutual Information Gain Maximization
    Wang, Xinzheng
    Guo, Bing
    Shen, Yan
    Zhou, Chimin
    Duan, Xuliang
    IEEE ACCESS, 2019, 7 : 151525 - 151538
  • [39] An improved feature selection method based on angle-guided multi-objective PSO and feature-label mutual information
    Han, Fei
    Wang, Tianyi
    Ling, Qinghua
    APPLIED INTELLIGENCE, 2023, 53 (03) : 3545 - 3562
  • [40] A Feature Subset Selection Method Based On High-Dimensional Mutual Information
    Zheng, Yun
    Kwoh, Chee Keong
    ENTROPY, 2011, 13 (04) : 860 - 901