Novel multi-label feature selection via label symmetric uncertainty correlation learning and feature redundancy evaluation

被引:67
作者
Dai, Jianhua [1 ,2 ]
Chen, Jiaolong [1 ,2 ]
Liu, Ye [1 ,2 ]
Hu, Hu [1 ]
机构
[1] Hunan Normal Univ, Hunan Prov Key Lab Intelligent Comp & Language In, Changsha 410081, Peoples R China
[2] Hunan Normal Univ, Coll Informat Sci & Engn, Changsha 410081, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-label feature selection; Symmetric uncertainty; Fuzzy mutual information; Feature redundancy evaluation; ATTRIBUTE SELECTION; MUTUAL INFORMATION; CLASSIFICATION; RECOGNITION; REDUCTION;
D O I
10.1016/j.knosys.2020.106342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-label data with high dimensionality, widely existed in the real world, bring many challenges to the applications of machine learning, pattern recognition and other fields. Scholars have proposed some multi-label feature selection methods from various aspects. However, there are few studies on the feature selection for multi-label data based on fuzzy mutual information, and most existing methods neglect the correlation between labels. In this study, we propose two novel multi-label feature selection approaches via label symmetric uncertainty correlation and feature redundancy evaluation. Firstly, we propose the concept of symmetric uncertainty correlation between labels via fuzzy mutual information, and design a label importance weight based on label symmetric uncertainty correlation learning. Further, we define a label similarity relation matrix on multi-label space via the label importance weight. Secondly, we define the symmetric uncertainty correlation between features and labels, and propose the first multi-label feature selection approach. Thirdly, considering the above-proposed method can only get a feature sequence and does not remove the redundancy features, we further propose an improved multi-label removing-redundancy feature selection approach through introducing feature redundancy evaluation. Finally, comprehensive experiments are executed to demonstrate the performance of our methods. The results illustrate that our study is better than other representative feature selection methods. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:16
相关论文
共 61 条
[41]  
Read J, 2009, LECT NOTES ARTIF INT, V5782, P254, DOI 10.1007/978-3-642-04174-7_17
[42]   BoosTexter: A boosting-based system for text categorization [J].
Schapire, RE ;
Singer, Y .
MACHINE LEARNING, 2000, 39 (2-3) :135-168
[43]  
Song X., 2020, IJCV, V128, P910, DOI DOI 10.1007/s11263-019-01287-w
[44]   Intuitionistic Fuzzy Rough Set-Based Granular Structures and Attribute Subset Selection [J].
Tan, Anhui ;
Wu, Wei-Zhi ;
Qian, Yuhua ;
Liang, Jiye ;
Chen, Jinkun ;
Li, Jinjin .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2019, 27 (03) :527-539
[45]  
Trohidis K., 1997, BLOOD, V90, P3438
[46]  
Tsoumakas G, 2010, DATA MINING AND KNOWLEDGE DISCOVERY HANDBOOK, SECOND EDITION, P667, DOI 10.1007/978-0-387-09823-4_34
[47]   Semantic annotation and retrieval of music and sound effects [J].
Turnbull, Douglas ;
Barrington, Luke ;
Torres, David ;
Lanckriet, Gert .
IEEE TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2008, 16 (02) :467-476
[48]   Multi-label classification using a fuzzy rough neighborhood consensus [J].
Vluymans, Sarah ;
Cornelis, Chris ;
Herrera, Francisco ;
Saeys, Yvan .
INFORMATION SCIENCES, 2018, 433 :96-114
[49]   Fuzzy rough set-based attribute reduction using distance measures [J].
Wang, Changzhong ;
Huang, Yang ;
Shao, Mingwen ;
Fan, Xiaodong .
KNOWLEDGE-BASED SYSTEMS, 2019, 164 :205-212
[50]   Feature subset selection based on fuzzy neighborhood rough sets [J].
Wang, Changzhong ;
Shao, Mingwen ;
He, Qiang ;
Qian, Yuhua ;
Qi, Yali .
KNOWLEDGE-BASED SYSTEMS, 2016, 111 :173-179