The logic transformations for reducing the complexity of the discernibility function-based attribute reduction problem

被引:3
作者
Hacibeyoglu, Mehmet [1 ]
Salman, Mohammad Shukri [2 ]
Selek, Murat [3 ]
Kahramanli, Sirzat [4 ]
机构
[1] Necmettin Erbakan Univ, Dept Comp Engn, Konya, Turkey
[2] Mevlana Univ, Dept Elect & Elect Engn, Konya, Turkey
[3] Selcuk Univ, Tech Vocat Sch Higher Educ, Konya, Turkey
[4] Mevlana Univ, Dept Comp Educ & Instruct Technol Teaching, Konya, Turkey
关键词
Attribute reduction; Bit-matrix partitioning; CNF to DNF conversion; Computational complexity; Discernibility function; Set cover; FEATURE-SELECTION; DIMENSIONALITY REDUCTION; PRIME IMPLICANTS; ROUGH SETS; CLASSIFICATION; MINIMIZATION; ALGORITHMS; MATRIX;
D O I
10.1007/s10115-015-0824-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The basic solution for locating an optimal reduct is to generate all possible reducts and select the one that best meets the given criterion. Since this problem is NP-hard, most attribute reduction algorithms use heuristics to find a single reduct with the risk to overlook for the best ones. There is a discernibility function (DF)-based approach that generates all reducts but may fail due to memory overflows even for datasets with dimensionality much below the medium. In this study, we show that the main shortcoming of this approach is its excessively high space complexity. To overcome this, we first represent a DF of attributes by a bit-matrix (BM). Second, we partition the BM into no more than sub-BMs (SBMs). Third, we convert each SBM into a subset of reducts by preventing the generation of redundant products, and finally, we unite the subsets into a complete set of reducts. Among the SBMs of a BM, the most complex one is the first SBM with a space complexity not greater than the square root of that of the original BM. The proposed algorithm converts such a SBM with attributes into the subset of reducts with the worst case space complexity of .
引用
收藏
页码:599 / 628
页数:30
相关论文
共 41 条
[1]  
BIEGANOWSKI J, 2005, SCHEDA INF, V14, P125
[2]   An efficient bit-based feature selection method [J].
Chen, Wei-Chou ;
Tseng, Shian-Shyong ;
Hong, Tzung-Pei .
EXPERT SYSTEMS WITH APPLICATIONS, 2008, 34 (04) :2858-2869
[3]   The Minimum Consistent Subset Cover Problem: A Minimization View of Data Mining [J].
Gao, Byron J. ;
Ester, Martin ;
Xiong, Hui ;
Cai, Jin-Yi ;
Schulte, Oliver .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2013, 25 (03) :690-703
[4]   A Hybrid Method for Fast Finding the Reduct with the Best Classification Accuracy [J].
Hacibeyoglu, Mehmet ;
Arslan, Ahmet ;
Kahramanli, Sirzat .
ADVANCES IN ELECTRICAL AND COMPUTER ENGINEERING, 2013, 13 (04) :57-64
[5]   A logic method for efficient reduction of the space complexity of the attribute reduction problem [J].
Hacibeyoglu, Mehmet ;
Basciftci, Fatih ;
Kahramanli, Sirzat .
TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2011, 19 (04) :643-656
[6]   Benchmarking attribute selection techniques for discrete class data mining [J].
Hall, MA ;
Holmes, G .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2003, 15 (06) :1437-1447
[7]   Feature Selection Based on Class-Dependent Densities for High-Dimensional Binary Data [J].
Javed, Kashif ;
Babri, Haroon A. ;
Saeed, Mehreen .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2012, 24 (03) :465-477
[8]   Semantics-preserving dimensionality reduction: Rough and fuzzy-rough-based approaches [J].
Jensen, R ;
Shen, Q .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2004, 16 (12) :1457-1471
[9]  
JENSEN R, 2007, ROUGH SET BASED ATTR
[10]  
Kahramanli S, 2011, INT J INNOV COMPUT I, V7, P2167