Multilabel Feature Selection Based on Fuzzy Mutual Information and Orthogonal Regression

被引:9
作者
Dai, Jianhua [1 ,2 ]
Liu, Qi [1 ,2 ]
Chen, Wenxiang [1 ,2 ]
Zhang, Chucai [1 ,2 ]
机构
[1] Hunan Normal Univ, Hunan Prov Key Lab Intelligent Comp & Language Inf, Changsha 410081, Peoples R China
[2] Hunan Normal Univ, Coll Informat Sci & Engn, Changsha 410081, Peoples R China
基金
中国国家自然科学基金;
关键词
Correlation; Mutual information; Feature extraction; Optimization; Redundancy; Fuzzy systems; Vectors; Feature selection; fuzzy rough sets; multilabel learning; mutual information; optimization objective; DEPENDENCY; REDUNDANCY; RELEVANCE;
D O I
10.1109/TFUZZ.2024.3415176
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the increase of high-dimensional multilabel data, multilabel feature selection (MFS) has received more and more widespread attention. Embedded feature selection methods have been widely studied due to their high efficiency and low computational cost. Fuzzy mutual information, as an effective tool for processing continuous features, is widely used in filter feature selection, which results in many repeated entropy calculations. Most of the existing multilabel embedded feature selection methods are based on least squares regression, which loses a lot of statistical and structural information. To solve the above-mentioned problems, we established an optimization framework based on fuzzy mutual information that considers global correlation to obtain the weight of each feature. Under this framework, many repeated entropy operations are avoided. Then, the weight of each feature is introduced into the orthogonal regression optimization framework as prior knowledge. Finally, two optimization frameworks are comprehensively considered for MFS. Furthermore, considering the characteristics of multilabel data, we extend the proposed method to feature-specific MFS. We conducted sufficient experiments to demonstrate the efficiency of our proposed method.
引用
收藏
页码:5136 / 5148
页数:13
相关论文
共 47 条
[1]   Feature selection using Joint Mutual Information Maximisation [J].
Bennasar, Mohamed ;
Hicks, Yulia ;
Setchi, Rossitza .
EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (22) :8520-8532
[2]   Learning multi-label scene classification [J].
Boutell, MR ;
Luo, JB ;
Shen, XP ;
Brown, CM .
PATTERN RECOGNITION, 2004, 37 (09) :1757-1771
[3]   The orthogonally constrained regression revisited [J].
Chu, MT ;
Trendafilov, NT .
JOURNAL OF COMPUTATIONAL AND GRAPHICAL STATISTICS, 2001, 10 (04) :746-771
[4]   Fuzzy Rough Attribute Reduction Based on Fuzzy Implication Granularity Information [J].
Dai, Jianhua ;
Zhu, Zhilin ;
Zou, Xiongtao .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (06) :3741-3752
[5]   Multi-label feature selection by strongly relevant label gain and label mutual aid [J].
Dai, Jianhua ;
Huang, Weiyi ;
Zhang, Chucai ;
Liu, Jie .
PATTERN RECOGNITION, 2024, 145
[6]   Multifuzzy b-Covering Approximation Spaces and Their Information Measures [J].
Dai, Jianhua ;
Zou, Xiongtao ;
Qian, Yuhua ;
Wang, Xizhao .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2023, 31 (03) :955-969
[7]   Feature selection via normative fuzzy information weight with application into tumor classification [J].
Dai, Jianhua ;
Chen, Jiaolong .
APPLIED SOFT COMPUTING, 2020, 92
[8]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[9]   Mutual information-based feature selection for multilabel classification [J].
Doquire, Gauthier ;
Verleysen, Michel .
NEUROCOMPUTING, 2013, 122 :148-155
[10]   MULTIPLE COMPARISONS AMONG MEANS [J].
DUNN, OJ .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1961, 56 (293) :52-&