Fuzzy Rough Based Feature Selection by Using Random Sampling

被引:2
作者
Wang Zhenlei [1 ]
Zhao Suyun [1 ]
Liu Yangming [1 ]
Chen Hong [1 ]
Li Cuiping [1 ]
Sun Xiran [1 ]
机构
[1] Renmin Univ China, Sch Informat, Beijing 100872, Peoples R China
来源
PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II | 2018年 / 11013卷
关键词
Randomly sampling; Fuzzy rough set; Attribute reduction; Maximum relevance; Minimum redundancy; INCREMENTAL APPROACH; ATTRIBUTE REDUCTION; APPROXIMATIONS; ALGORITHMS;
D O I
10.1007/978-3-319-97310-4_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Feature selection, i.e., Attribute reduction, is one of the most important applications of fuzzy rough set theory. The application of attribute reduction based on fuzzy rough set is inefficient or even unfeasible on large scale data. Considering the random sampling technique is an effective method to statistically reduce the calculation on large scale data, we introduce it into the fuzzy rough based feature selection algorithm. This paper thus proposes a random reduction algorithm based on random sampling. The main contribution of this paper is the introduction of the idea of random sampling in the selection of attributes based on minimum redundancy and maximum correlation. First, in each iteration the significance of attribute is not computed on all the objects in the whole datasets, but on part of randomly selected objects. By this way, the maximum relevant attribute is chosen on the condition of less calculation. Secondly, in the process of choosing attribute in each iteration, the sample is different so as to select the minimum redundancy attribute. Finally, the experimental results show that the reduction algorithm can obviously reduce the running time of the reduction algorithm on the condition of limited classification accuracy loss.
引用
收藏
页码:91 / 99
页数:9
相关论文
共 27 条
  • [1] Anagnostopoulos E, 2016, COMPUT SCI
  • [2] [Anonymous], 2008, Advances in Neural Information Processing Systems, DOI DOI 10.7751/mitpress/8996.003.0015
  • [3] Selection of relevant features and examples in machine learning
    Blum, AL
    Langley, P
    [J]. ARTIFICIAL INTELLIGENCE, 1997, 97 (1-2) : 245 - 271
  • [4] A Rough-Set-Based Incremental Approach for Updating Approximations under Dynamic Maintenance Environments
    Chen, Hongmei
    Li, Tianrui
    Ruan, Da
    Lin, Jianhui
    Hu, Chengxiang
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2013, 25 (02) : 274 - 284
  • [5] NEAREST NEIGHBOR PATTERN CLASSIFICATION
    COVER, TM
    HART, PE
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) : 21 - +
  • [6] Dai H, 2017, ANALYSIS FOR TIME-TO-EVENT DATA UNDER CENSORING AND TRUNCATION, P1
  • [7] Rough approximations on a complete completely distributive lattice with applications to generalized rough sets
    Degang, Chen
    Wenxiu, Zhang
    Yeung, Daniel
    Tsang, E. C. C.
    [J]. INFORMATION SCIENCES, 2006, 176 (13) : 1829 - 1848
  • [8] ROUGH FUZZY-SETS AND FUZZY ROUGH SETS
    DUBOIS, D
    PRADE, H
    [J]. INTERNATIONAL JOURNAL OF GENERAL SYSTEMS, 1990, 17 (2-3) : 191 - 209
  • [9] Filtering algorithms for global chance constraints
    Hnich, Brahim
    Rossi, Roberto
    Tarim, S. Armagan
    Prestwich, Steven
    [J]. ARTIFICIAL INTELLIGENCE, 2012, 189 : 69 - 94
  • [10] Information-preserving hybrid data reduction based on fuzzy-rough techniques
    Hu, QH
    Yu, DR
    Xie, ZX
    [J]. PATTERN RECOGNITION LETTERS, 2006, 27 (05) : 414 - 423