An efficient accelerator for attribute reduction from incomplete data in rough set framework

被引:191
作者
Qian, Yuhua [1 ,2 ]
Liang, Jiye [1 ]
Pedrycz, Witold [3 ]
Dang, Chuangyin [2 ]
机构
[1] Shanxi Univ, Key Lab Computat Intelligence & Chinese Informat, Minist Educ, Taiyuan 030006, Shanxi, Peoples R China
[2] City Univ Hong Kong, Dept Mfg Engn & Engn Management, Hong Kong, Hong Kong, Peoples R China
[3] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB, Canada
关键词
Feature selection; Rough set theory; Incomplete information systems; Positive approximation; Granular computing; FEATURE-SELECTION; KNOWLEDGE REDUCTION; INFORMATION; GRANULATION; ENTROPY; DIMENSIONALITY; UNCERTAINTY; SYSTEMS;
D O I
10.1016/j.patcog.2011.02.020
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection (attribute reduction) from large-scale incomplete data is a challenging problem in areas such as pattern recognition, machine learning and data mining. In rough set theory, feature selection from incomplete data aims to retain the discriminatory power of original features. To address this issue, many feature selection algorithms have been proposed, however, these algorithms are often computationally time-consuming. To overcome this shortcoming, we introduce in this paper a theoretic framework based on rough set theory, which is called positive approximation and can be used to accelerate a heuristic process for feature selection from incomplete data. As an application of the proposed accelerator, a general feature selection algorithm is designed. By integrating the accelerator into a heuristic algorithm, we obtain several modified representative heuristic feature selection algorithms in rough set theory. Experiments show that these modified algorithms outperform their original counterparts. It is worth noting that the performance of the modified algorithms becomes more visible when dealing with larger data sets. (C) 2011 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1658 / 1670
页数:13
相关论文
共 46 条
[1]   On the compact computational domain of fuzzy-rough sets [J].
Bhatt, RB ;
Gopal, M .
PATTERN RECOGNITION LETTERS, 2005, 26 (11) :1632-1640
[2]   On fuzzy-rough sets approach to feature selection [J].
Bhatt, RB ;
Gopal, M .
PATTERN RECOGNITION LETTERS, 2005, 26 (07) :965-975
[3]   Global discretization of continuous attributes as preprocessing for machine learning [J].
Chmielewski, MR ;
GrzymalaBusse, JW .
INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 1996, 15 (04) :319-331
[4]   Consistency-based search in feature selection [J].
Dash, M ;
Liu, HA .
ARTIFICIAL INTELLIGENCE, 2003, 151 (1-2) :155-176
[5]  
Guyon I., 2003, J MACH LEARN RES, V3, P1157
[6]   Fuzzy probabilistic approximation spaces and their information measures [J].
Hu, QH ;
Yu, DR ;
Xie, ZX ;
Liu, JF .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2006, 14 (02) :191-201
[7]   Information-preserving hybrid data reduction based on fuzzy-rough techniques [J].
Hu, QH ;
Yu, DR ;
Xie, ZX .
PATTERN RECOGNITION LETTERS, 2006, 27 (05) :414-423
[8]   Hybrid attribute reduction based on a novel fuzzy-rough model and information granulation [J].
Hu, Qinghua ;
Xie, Zongxia ;
Yu, Daren .
PATTERN RECOGNITION, 2007, 40 (12) :3509-3521
[9]   LEARNING IN RELATIONAL DATABASES - A ROUGH SET APPROACH [J].
HU, XH ;
CERCONE, N .
COMPUTATIONAL INTELLIGENCE, 1995, 11 (02) :323-338
[10]  
[黄兵 Huang Bing], 2005, [系统工程理论与实践, Systems Engineering-Theory & Practice], V25, P55