System Entropy and Its Application in Feature Selection

被引:2
作者
ZHAO JunWU ZhongfuLI Hua Institute of Computer Science and Technology Chongqing University of Posts and Telecommunications Chongqing PRChinaCollege of Computer Science and Engineering Chongqing University Chongqing PRChina [1 ,2 ,2 ,2 ,1 ,400065 ,2 ,400044 ]
机构
关键词
feature selection; system entropy; rough set theory; data mining;
D O I
暂无
中图分类号
TP391.4 [模式识别与装置];
学科分类号
0811 ; 081101 ; 081104 ; 1405 ;
摘要
Feature selection is always an important issue in the research on data mining technologies. However, the problem of optimal feature selection is NP hard. Therefore, heuristic approaches are more practical to actual learning systems. Usually, that kind of algorithm selects features with the help of a heuristic metric compactum to measure the relative importance of features in a learning system. Here a new notion of ‘system entropy’ is described in terms of rough set theory, and then some of its algebraic characteristics are studied. After its intrinsic value biase is effectively counteracted, the system entropy is applied in BSE, a new heuristic algorithm for feature selection. BSE is efficient, whose time complexity is lower than that of analogous algorithms; BSE is also effective, which can produce the optimal results in the mini-feature biased sense from varieties of learning systems. Besides, BSE is tolerant and also flexible to the inconsistency of a learning system, consequently able to elegantly handle data noise in the learning system.
引用
收藏
页码:100 / 105
页数:6
相关论文
共 16 条
[11]  
Greedy attribute selection. CARUANA R,FREITAG D. Proceedings of the 11th International Conference on Machine Learning . 1994
[12]  
The mathematical theory of communication. SHANNON C L. The Bell System Technical Journal . 1948
[13]  
Machinelearningrepository. http:∥www.ics.uci.edu/~mlearn/MLRepository.html .
[14]  
An information repre-sentation of the concepts and operation in rough set theory. MIAO D Q,WANG Jue. Journal of Software . 1999
[15]  
Decision table reduction based on conditional information entropy. WANG G Y,YU H,YANG D C. Chinese Journal of Computer . 2002
[16]  
Toward optimal feature selection. KOLLER D,SAHAMI M. Proceedings of the 13th International Conf on M L . 1996