iSOUP-SymRF: Symbolic feature ranking with random forests in online multi-target regression and multi-label classification

被引:0
作者
Osojnik, Aljaz [1 ]
Panov, Pance [1 ,2 ]
Dzeroski, Saso [1 ,2 ]
机构
[1] Jozef Stefan Inst, Dept Knowledge Technol, Jamova 39, Ljubljana, Slovenia
[2] Jozef Stefan Int Postgrad Sch, Jamova 39, Ljubljana, Slovenia
基金
欧盟地平线“2020”;
关键词
Online learning; Feature ranking; Multi-target regression; Multi-label classification; FEATURE-SELECTION;
D O I
10.1007/s10994-024-06718-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The task of feature ranking has received considerable attention across various predictive modelling tasks in the batch learning scenario, but not in the online learning setting. Available methods that estimate feature importances on data streams have so far predominantly focused on ranking the features for the tasks of classification and occasionally multi-label classification. We propose a novel online feature ranking method for online multi-target regression iSOUP-SymRF, which estimates feature importance scores based on the positions at which a feature appears in the trees of a random forest of iSOUP-Trees, and additionally extend it to task of online feature ranking for multi-label classification. By utilizing iSOUP-Trees, which can address multiple structured output prediction tasks on data streams, iSOUP-SymRF promises feature ranking across a variety of online structured output prediction tasks. We examine the ranking convergence of iSOUP-SymRF in terms of the methods' parameters, the size of the ensemble and the number of selected features, as well as their stability under different random seeds. Furthermore, to show the utility of iSOUP-SymRF and its rankings we use them in conjunction with two state-of-the-art online multi-target regression and multi-label classification methods, iSOUP-Tree and AMRules, and analyze the impact of adding features according to the rankings obtained from iSOUP-SymRF.
引用
收藏
页数:24
相关论文
共 50 条
[41]   Feature ranking for enhancing boosting-based multi-label text categorization [J].
Al-Salemi, Bassam ;
Ayob, Masri ;
Noah, Shahrul Azman Mohd .
EXPERT SYSTEMS WITH APPLICATIONS, 2018, 113 :531-543
[42]   Multi-task Joint Feature Selection for Multi-label Classification [J].
HE Zhifen ;
YANG Ming ;
LIU Huidong .
Chinese Journal of Electronics, 2015, 24 (02) :281-287
[43]   Multi-task Joint Feature Selection for Multi-label Classification [J].
He Zhifen ;
Yang Ming ;
Liu Huidong .
CHINESE JOURNAL OF ELECTRONICS, 2015, 24 (02) :281-287
[44]   Feature selection for multi-label naive Bayes classification [J].
Zhang, Min-Ling ;
Pena, Jose M. ;
Robles, Victor .
INFORMATION SCIENCES, 2009, 179 (19) :3218-3229
[45]   Multi-label text classification with an ensemble feature space [J].
Tandon, Kushagri ;
Chatterjee, Niladri .
JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (05) :4425-4436
[46]   Dynamic feature weighting for multi-label classification problems [J].
Maryam Dialameh ;
Ali Hamzeh .
Progress in Artificial Intelligence, 2021, 10 :283-295
[47]   Categorizing feature selection methods for multi-label classification [J].
Pereira, Rafael B. ;
Plastino, Alexandre ;
Zadrozny, Bianca ;
Merschmann, Luiz H. C. .
ARTIFICIAL INTELLIGENCE REVIEW, 2018, 49 (01) :57-78
[48]   Categorizing feature selection methods for multi-label classification [J].
Rafael B. Pereira ;
Alexandre Plastino ;
Bianca Zadrozny ;
Luiz H. C. Merschmann .
Artificial Intelligence Review, 2018, 49 :57-78
[49]   Multi-label Classification using Random Walk with Restart [J].
Liu, Jinhong ;
Yang, Juan .
2017 INTERNATIONAL CONFERENCE ON CYBER-ENABLED DISTRIBUTED COMPUTING AND KNOWLEDGE DISCOVERY (CYBERC), 2017, :206-212
[50]   Optimizing margin distribution for online multi-label classification [J].
Zhai, Tingting ;
Hu, Kunyong .
EVOLVING SYSTEMS, 2024, 15 (03) :1033-1042