Multi-Label Learning with Distribution Matching Ensemble: An Adaptive and Just-In-Time Weighted Ensemble Learning Algorithm for Classifying a Nonstationary Online Multi-Label Data Stream

被引:1
|
作者
Shen, Chao [1 ]
Liu, Bingyu [1 ]
Shao, Changbin [1 ]
Yang, Xibei [1 ]
Xu, Sen [2 ]
Zhu, Changming [3 ]
Yu, Hualong [1 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Comp, Zhenjiang 212100, Peoples R China
[2] Yancheng Inst Technol, Sch Informat Technol, Yancheng 224051, Peoples R China
[3] Minzu Univ China, Key Lab Ethn language Intelligent Anal & Secur Gov, Beijing 100081, Peoples R China
来源
SYMMETRY-BASEL | 2025年 / 17卷 / 02期
基金
中国国家自然科学基金;
关键词
multi-label data stream; adaptive weighted ensemble; concept drift; distribution matching; Gaussian mixture model; Kullback-Leibler divergence; label distribution drift detection; CONCEPT DRIFT; CLASSIFICATION; MACHINE;
D O I
10.3390/sym17020182
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Learning from a nonstationary data stream is challenging, as a data stream is generally considered to be endless, and the learning model is required to be constantly amended for adapting the shifting data distributions. When it meets multi-label data, the challenge would be further intensified. In this study, an adaptive online weighted multi-label ensemble learning algorithm called MLDME (multi-label learning with distribution matching ensemble) is proposed. It simultaneously calculates both the feature matching level and label matching level between any one reserved data block and the new received data block, further providing an adaptive decision weight assignment for ensemble classifiers based on their distribution similarities. Specifically, MLDME abandons the most commonly used but not totally correct underlying hypothesis that in a data stream, each data block always has the most approximate distribution with that emerging after it; thus, MLDME could provide a just-in-time decision for the new received data block. In addition, to avoid an infinite extension of ensemble classifiers, we use a fixed-size buffer to store them and design three different dynamic classifier updating rules. Experimental results for nine synthetic and three real-world multi-label nonstationary data streams indicate that the proposed MLDME algorithm is superior to some popular and state-of-the-art online learning paradigms and algorithms, including two specifically designed ones for classifying a nonstationary multi-label data stream.
引用
收藏
页数:25
相关论文
共 50 条
  • [31] Three-Way Decisions Based Multi-label Learning Algorithm with Label Dependency
    Li, Feng
    Miao, Duoqian
    Zhang, Wei
    ROUGH SETS, (IJCRS 2016), 2016, 9920 : 240 - 249
  • [32] Mutual information-based label distribution feature selection for multi-label learning
    Qian, Wenbin
    Huang, Jintao
    Wang, Yinglong
    Shu, Wenhao
    KNOWLEDGE-BASED SYSTEMS, 2020, 195
  • [33] Mining Multi-label Concept-Drifting Data Streams Using Dynamic Classifier Ensemble
    Qu, Wei
    Zhang, Yang
    Zhu, Junping
    Qiu, Qiang
    ADVANCES IN MACHINE LEARNING, PROCEEDINGS, 2009, 5828 : 308 - 321
  • [34] Instance selection for multi-label learning based on a scalable evolutionary algorithm
    Romero del Castillo, Juan Antonio
    Ortiz-Boyer, Domingo
    Garcia-Pedrajas, Nicolas
    21ST IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS ICDMW 2021, 2021, : 843 - 851
  • [35] A novel approach for learning label correlation with application to feature selection of multi-label data
    Che, Xiaoya
    Chen, Degang
    Mi, Jusheng
    INFORMATION SCIENCES, 2020, 512 (512) : 795 - 812
  • [36] Efficient Class Incremental Learning for Multi-label Classification of Evolving Data Streams
    Shi, Zhongwei
    Wen, Yimin
    Xue, Yun
    Cai, Guoyong
    PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 2093 - 2099
  • [37] Imbalanced and missing multi-label data learning with global and local structure
    Su, Xinpei
    Xu, Yitian
    INFORMATION SCIENCES, 2024, 677
  • [38] An Improved Multi-Label Learning Method with ELM-RBF and a Synergistic Adaptive Genetic Algorithm
    Zhang, Dezheng
    Li, Peng
    Wulamu, Aziguli
    ALGORITHMS, 2022, 15 (06)
  • [39] Boosting label weighted extreme learning machine for classifying multi -label imbalanced data
    Cheng, Ke
    Gao, Shang
    Dong, Wenlu
    Yang, Xibei
    Wang, Qi
    Yu, Hualong
    NEUROCOMPUTING, 2020, 403 (403) : 360 - 370
  • [40] ML-KnockoffGAN: Deep online feature selection for multi-label learning
    Paul, Dipanjyoti
    Bardhan, Snigdha
    Saha, Sriparna
    Mathew, Jimson
    KNOWLEDGE-BASED SYSTEMS, 2023, 271