Fast and Robust Attribute Reduction Based on the Separability in Fuzzy Decision Systems

被引:73
作者
Hu, Meng [1 ]
Tsang, Eric C. C. [1 ]
Guo, Yanting [1 ]
Xu, Weihua [2 ]
机构
[1] Macau Univ Sci & Technol, Fac Informat Technol, Macau, Peoples R China
[2] Southwest Univ, Coll Artificial Intelligence, Chongqing 400715, Peoples R China
关键词
Entropy; Kernel; Task analysis; Rough sets; Machine learning; Redundancy; Mutual information; Attribute reduction; fuzzy decision systems; fuzzy membership; separability; FEATURE-SELECTION; MUTUAL INFORMATION; ENTROPY; CLASSIFICATION; UNCERTAINTY; CLASSIFIERS; SETS;
D O I
10.1109/TCYB.2020.3040803
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Attribute reduction is one of the most important preprocessing steps in machine learning and data mining. As a key step of attribute reduction, attribute evaluation directly affects classification performance, search time, and stopping criterion. The existing evaluation functions are greatly dependent on the relationship between objects, which makes its computational time and space more costly. To solve this problem, we propose a novel separability-based evaluation function and reduction method by using the relationship between objects and decision categories directly. The degree of aggregation (DA) of intraclass objects and the degree of dispersion (DD) of between-class objects are first defined to measure the significance of an attribute subset. Then, the separability of attribute subsets is defined by DA and DD in fuzzy decision systems, and we design a sequentially forward selection based on the separability (SFSS) algorithm to select attributes. Furthermore, a postpruning strategy is introduced to prevent overfitting and determine a termination parameter. Finally, the SFSS algorithm is compared with some typical reduction algorithms using some public datasets from UCI and ELVIRA Biomedical repositories. The interpretability of SFSS is directly presented by the performance on MNIST handwritten digits. The experimental comparisons show that SFSS is fast and robust, which has higher classification accuracy and compression ratio, with extremely low computational time.
引用
收藏
页码:5559 / 5572
页数:14
相关论文
共 48 条
[11]  
Dua D, 2019, UCI MACHINE LEARNING
[12]   ROUGH FUZZY-SETS AND FUZZY ROUGH SETS [J].
DUBOIS, D ;
PRADE, H .
INTERNATIONAL JOURNAL OF GENERAL SYSTEMS, 1990, 17 (2-3) :191-209
[13]   FUZZY-SETS IN APPROXIMATE REASONING .1. INFERENCE WITH POSSIBILITY DISTRIBUTIONS [J].
DUBOIS, D ;
PRADE, H .
FUZZY SETS AND SYSTEMS, 1991, 40 (01) :143-202
[14]   A comparative analysis of methods for pruning decision trees [J].
Esposito, F ;
Malerba, D ;
Semeraro, G .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (05) :476-491
[15]   The use of multiple measurements in taxonomic problems [J].
Fisher, RA .
ANNALS OF EUGENICS, 1936, 7 :179-188
[16]   A comparison of alternative tests of significance for the problem of m rankings [J].
Friedman, M .
ANNALS OF MATHEMATICAL STATISTICS, 1940, 11 :86-92
[17]   Incremental updating approximations for double-quantitative decision-theoretic rough sets with the variation of objects [J].
Guo, Yanting ;
Tsang, Eric C. C. ;
Hu, Meng ;
Lin, Xuxin ;
Chen, Degang ;
Xu, Weihua ;
Sang, Binbin .
KNOWLEDGE-BASED SYSTEMS, 2020, 189
[18]   Adaptive weighted generalized multi-granulation interval-valued decision-theoretic rough sets [J].
Guo, Yanting ;
Tsang, Eric C. C. ;
Xu, Weihua ;
Chen, Degang .
KNOWLEDGE-BASED SYSTEMS, 2020, 187
[19]   Local logical disjunction double-quantitative rough sets [J].
Guo, Yanting ;
Tsang, Eric C. C. ;
Xu, Weihua ;
Chen, Degang .
INFORMATION SCIENCES, 2019, 500 :87-112
[20]   A reformulation of entropy in the presence of indistinguishability operators [J].
Hernández, E ;
Recasens, J .
FUZZY SETS AND SYSTEMS, 2002, 128 (02) :185-196