Incremental feature selection based on rough set in dynamic incomplete data

被引:115
作者
Shu, Wenhao [1 ]
Shen, Hong [2 ,3 ]
机构
[1] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing, Peoples R China
[2] Sun Yat Sen Univ, Sch Informat Sci & Technol, Guangzhou, Guangdong, Peoples R China
[3] Univ Adelaide, Sch Comp Sci, Adelaide, SA 5005, Australia
基金
美国国家科学基金会;
关键词
Feature selection; Positive region; Rough set theory; Dynamic incomplete data; ATTRIBUTE REDUCTION; INFORMATION; SYSTEMS; INDISCERNIBILITY; APPROXIMATION; VALUES;
D O I
10.1016/j.patcog.2014.06.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection plays a vital role in many areas of pattern recognition and data mining. The effective computation of feature selection is important for improving the classification performance. In rough set theory, many feature selection algorithms have been proposed to process static incomplete data. However, feature values in an incomplete data set may vary dynamically in real-world applications. For such dynamic incomplete data, a classic (non-incremental) approach of feature selection is usually computationally time-consuming. To overcome this disadvantage, we propose an incremental approach for feature selection, which can accelerate the feature selection process in dynamic incomplete data. We firstly employ an incremental manner to compute the new positive region when feature values with respect to an object set vary dynamically. Based on the calculated positive region, two efficient incremental feature selection algorithms are developed respectively for single object and multiple objects with varying feature values. Then we conduct a series of experiments with 12 UCI real data sets to evaluate the efficiency and effectiveness of our proposed algorithms. The experimental results show that the proposed algorithms compare favorably with that of applying the existing non-incremental methods. (C) 2014 Elsevier Ltd. All rights reserved.
引用
收藏
页码:3890 / 3906
页数:17
相关论文
共 59 条
[1]  
[Anonymous], ROUGH SETS KNOWLEDGE
[2]  
[Anonymous], 1992, Intelligent Decision Support. Handbook of Applications and Advances of the Rough Sets Theory, DOI DOI 10.1007/978-94-015-7975-9_21
[3]   Sample Pair Selection for Attribute Reduction with Rough Set [J].
Chen, Degang ;
Zhao, Suyun ;
Zhang, Lei ;
Yang, Yongping ;
Zhang, Xiao .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2012, 24 (11) :2080-2093
[4]   A Rough Set Based Dynamic Maintenance Approach for Approximations in Coarsening and Refining Attribute Values [J].
Chen, Hongmei ;
Li, Tianrui ;
Qiao, Shaojie ;
Ruan, Da .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2010, 25 (10) :1005-1026
[5]   The incremental method for fast computing the rough fuzzy approximations [J].
Cheng, Yi .
DATA & KNOWLEDGE ENGINEERING, 2011, 70 (01) :84-100
[6]   Estimating optimal feature subsets using efficient estimation of high-dimensional mutual information [J].
Chow, TWS ;
Huang, D .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 2005, 16 (01) :213-224
[7]   Consistency-based search in feature selection [J].
Dash, M ;
Liu, HA .
ARTIFICIAL INTELLIGENCE, 2003, 151 (1-2) :155-176
[8]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[9]   Feature Selection With Harmony Search [J].
Diao, Ren ;
Shen, Qiang .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2012, 42 (06) :1509-1523
[10]   A comparison of alternative tests of significance for the problem of m rankings [J].
Friedman, M .
ANNALS OF MATHEMATICAL STATISTICS, 1940, 11 :86-92