Multi-Label Learning with Distribution Matching Ensemble: An Adaptive and Just-In-Time Weighted Ensemble Learning Algorithm for Classifying a Nonstationary Online Multi-Label Data Stream

被引:1
|
作者
Shen, Chao [1 ]
Liu, Bingyu [1 ]
Shao, Changbin [1 ]
Yang, Xibei [1 ]
Xu, Sen [2 ]
Zhu, Changming [3 ]
Yu, Hualong [1 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Comp, Zhenjiang 212100, Peoples R China
[2] Yancheng Inst Technol, Sch Informat Technol, Yancheng 224051, Peoples R China
[3] Minzu Univ China, Key Lab Ethn language Intelligent Anal & Secur Gov, Beijing 100081, Peoples R China
来源
SYMMETRY-BASEL | 2025年 / 17卷 / 02期
基金
中国国家自然科学基金;
关键词
multi-label data stream; adaptive weighted ensemble; concept drift; distribution matching; Gaussian mixture model; Kullback-Leibler divergence; label distribution drift detection; CONCEPT DRIFT; CLASSIFICATION; MACHINE;
D O I
10.3390/sym17020182
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Learning from a nonstationary data stream is challenging, as a data stream is generally considered to be endless, and the learning model is required to be constantly amended for adapting the shifting data distributions. When it meets multi-label data, the challenge would be further intensified. In this study, an adaptive online weighted multi-label ensemble learning algorithm called MLDME (multi-label learning with distribution matching ensemble) is proposed. It simultaneously calculates both the feature matching level and label matching level between any one reserved data block and the new received data block, further providing an adaptive decision weight assignment for ensemble classifiers based on their distribution similarities. Specifically, MLDME abandons the most commonly used but not totally correct underlying hypothesis that in a data stream, each data block always has the most approximate distribution with that emerging after it; thus, MLDME could provide a just-in-time decision for the new received data block. In addition, to avoid an infinite extension of ensemble classifiers, we use a fixed-size buffer to store them and design three different dynamic classifier updating rules. Experimental results for nine synthetic and three real-world multi-label nonstationary data streams indicate that the proposed MLDME algorithm is superior to some popular and state-of-the-art online learning paradigms and algorithms, including two specifically designed ones for classifying a nonstationary multi-label data stream.
引用
收藏
页数:25
相关论文
共 50 条
  • [41] A Novel Online Real-time Classifier for Multi-label Data Streams
    Venkatesan, Rajasekar
    Er, Meng Joo
    Wu, Shiqian
    Pratama, Mahardhika
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 1833 - 1840
  • [42] Transductive Multi-Instance Multi-Label learning algorithm with application to automatic image annotation
    Feng, Songhe
    Xu, De
    EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (01) : 661 - 670
  • [43] Cost Sensitive Ranking Support Vector Machine for Multi-label Data Learning
    Cao, Peng
    Liu, Xiaoli
    Zhao, Dazhe
    Zaiane, Osmar
    PROCEEDINGS OF THE 16TH INTERNATIONAL CONFERENCE ON HYBRID INTELLIGENT SYSTEMS (HIS 2016), 2017, 552 : 244 - 255
  • [44] Learning from multi-label data with interactivity constraints: An extensive experimental study
    Nair-Benrekia, Noureddine-Yassine
    Kuntz, Pascale
    Meyer, Frank
    EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (13) : 5723 - 5736
  • [45] Performance comparison of multi-label learning algorithms on clinical data for chronic diseases
    Zufferey, Damien
    Hofer, Thomas
    Hennebert, Jean
    Schumacher, Michael
    Ingold, Rolf
    Bromuri, Stefano
    COMPUTERS IN BIOLOGY AND MEDICINE, 2015, 65 : 34 - 43
  • [46] Adaptive feature selection in PET scans based on shared information and multi-label learning
    Sbei, Arafet
    ElBedoui, Khaoula
    Barhoumi, Walid
    Maktouf, Chokri
    VISUAL COMPUTER, 2022, 38 (01) : 257 - 277
  • [47] A novel ensemble over-sampling approach based Chebyshev inequality for imbalanced multi-label data
    Ren, Weishuo
    Zheng, Yifeng
    Zhang, Wenjie
    Qing, Depeng
    Zeng, Xianlong
    Li, Guohe
    NEUROCOMPUTING, 2025, 612
  • [48] A k-Nearest Neighbor Based Algorithm for Multi-Instance Multi-Label Active Learning
    Ruiz, Adrian T.
    Thiam, Patrick
    Schwenker, Friedhelm
    Palm, Guenther
    ARTIFICIAL NEURAL NETWORKS IN PATTERN RECOGNITION, ANNPR 2018, 2018, 11081 : 139 - 151
  • [49] A relative labeling importance estimation algorithm based on global-local label correlations for multi-label learning
    Liu, Yilu
    Cao, Fuyuan
    APPLIED INTELLIGENCE, 2023, 53 (05) : 4940 - 4958
  • [50] Self-adjusting k nearest neighbors for continual learning from multi-label drifting data streams
    Roseberry, Martha
    Krawczyk, Bartosz
    Djenouri, Youcef
    Cano, Alberto
    NEUROCOMPUTING, 2021, 442 : 10 - 25