Multi-level rough set reduction for decision rule mining

被引:17
作者
Ye, Mingquan [1 ,2 ]
Wu, Xindong [1 ]
Hu, Xuegang [1 ]
Hu, Donghui [1 ]
机构
[1] Hefei Univ Technol, Dept Comp Sci, Hefei 230009, Peoples R China
[2] Wannan Med Coll, Dept Comp Sci, Wuhu 241002, Peoples R China
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Rough set theory; Multi-level data mining; Attribute value taxonomy; Data generalization; Concept level; ATTRIBUTE REDUCTION; INFORMATION; APPROXIMATION; ACQUISITION; EXTRACTION; TAXONOMIES; NETWORKS; VALUES; TABLES;
D O I
10.1007/s10489-013-0439-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most previous studies on rough sets focused on attribute reduction and decision rule mining on a single concept level. Data with attribute value taxonomies (AVTs) are, however, commonly seen in real-world applications. In this paper, we extend Pawlak's rough set model, and propose a novel multi-level rough set model (MLRS) based on AVTs and a full-subtree generalization scheme. Paralleling with Pawlak's rough set model, some conclusions related to the MLRS are given. Meanwhile, a novel concept of cut reduction based on MLRS is presented. A cut reduction can induce the most abstract multi-level decision table with the same classification ability on the raw decision table, and no other multi-level decision table exists that is more abstract. Furthermore, the relationships between attribute reduction in Pawlak's rough set model and cut reduction in MLRS are discussed. We also prove that the problem of cut reduction generation is NP-hard, and develop a heuristic algorithm named CRTDR for computing the cut reduction. Finally, an approach named RMTDR for mining multi-level decision rule is provided. It can mine decision rules from different concept levels. Example analysis and comparative experiments show that the proposed methods are efficient and effective in handling the problems where data is associated with AVTs.
引用
收藏
页码:642 / 658
页数:17
相关论文
共 44 条
[1]  
[Anonymous], 1985, Bull. Pol. Acad. Sci. Math
[2]   Maintenance of approximations in incomplete ordered decision systems while attribute values coarsening or refining [J].
Chen, Hongmei ;
Li, Tianrui ;
Ruan, Da .
KNOWLEDGE-BASED SYSTEMS, 2012, 31 :140-161
[3]   From data to global generalized knowledge [J].
Chen, Yen-Liang ;
Wu, Yu-Ying ;
Chang, Ray-I .
DECISION SUPPORT SYSTEMS, 2012, 52 (02) :295-307
[4]   Cluster-based instance selection for machine classification [J].
Czarnowski, Ireneusz .
KNOWLEDGE AND INFORMATION SYSTEMS, 2012, 30 (01) :113-133
[5]   Learning structured bayesian networks: Combining abstraction hierarchies and tree-structured conditional probability tables [J].
Desjardins, Marie ;
Rathod, Priyang ;
Getoor, Lise .
COMPUTATIONAL INTELLIGENCE, 2008, 24 (01) :1-22
[6]   Improving classification performance of Support Vector Machine by genetically optimising kernel shape and hyper-parameters [J].
Diosan, Laura ;
Rogozan, Alexandrina ;
Pecuchet, Jean-Pierre .
APPLIED INTELLIGENCE, 2012, 36 (02) :280-294
[7]   A vague-rough set approach for uncertain knowledge acquisition [J].
Feng, Lin ;
Li, Tianrui ;
Ruan, Da ;
Gou, Shirong .
KNOWLEDGE-BASED SYSTEMS, 2011, 24 (06) :837-843
[8]   Hierarchical decision rules mining [J].
Feng, Qinrong ;
Miao, Duoqian ;
Cheng, Yi .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (03) :2081-2091
[9]   Feature subset selection wrapper based on mutual information and rough sets [J].
Foithong, Sombut ;
Pinngern, Ouen ;
Attachoo, Boonwat .
EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (01) :574-584
[10]   Rough computational methods for information systems [J].
Guan, JW ;
Bell, DA .
ARTIFICIAL INTELLIGENCE, 1998, 105 (1-2) :77-103