Multi-label borderline oversampling technique

被引:10
|
作者
Teng, Zeyu [1 ]
Cao, Peng [2 ,3 ]
Huang, Min [1 ]
Gao, Zheming [1 ]
Wang, Xingwei [2 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang 110819, Liaoning, Peoples R China
[2] Northeastern Univ, Coll Comp Sci & Engn, Shenyang 110169, Liaoning, Peoples R China
[3] Northeastern Univ, Key Lab Intelligent Comp Med Image, Minist Educ, Shenyang 110169, Liaoning, Peoples R China
关键词
Multi-label learning; Class imbalance; Borderline sample; Oversampling; CLASSIFICATION; IMBALANCE; RANKING; MACHINE; SMOTE;
D O I
10.1016/j.patcog.2023.109953
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Class imbalance problem commonly exists in multi-label classification (MLC) tasks. It has non-negligible im-pacts on the classifier performance and has drawn extensive attention in recent years. Borderline oversampling has been widely used in single-label learning as a competitive technique in dealing with class imbalance. Nevertheless, the borderline samples in multi-label data sets (MLDs) have not been studied. Hence, this paper deeply discussed the borderline samples in MLDs and found they have different neighboring relationships with class borders, which makes their roles different in the classifier training. For that, they are divided into two types named the self-borderline samples and the cross-borderline samples. Further, a novel MLDs resampling approach called Multi-Label Borderline Oversampling Technique (MLBOTE) is proposed for multi -label imbalanced learning. MLBOTE identifies three types of seed samples, including interior, self-borderline, and cross-borderline samples, and different oversampling mechanisms are designed for them, respectively. Meanwhile, it regards not only the minority classes but also the classes suffering from one-vs-rest imbalance as those in need of oversampling. Experiments on eight data sets with nine MLC algorithms and three base classifiers are carried out to compare MLBOTE with some state-of-art MLDs resampling techniques. The results show MLBOTE outperforms other methods in various scenarios.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] Multi-Label Learning with Label Enhancement
    Shao, Ruifeng
    Xu, Ning
    Geng, Xin
    2018 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2018, : 437 - 446
  • [32] Addressing class-imbalance in multi-label learning via two-stage multi-label hypernetwork
    Sun, Kai Wei
    Lee, Chong Ho
    NEUROCOMPUTING, 2017, 266 : 375 - 389
  • [33] Towards Class-Imbalance Aware Multi-Label Learning
    Zhang, Min-Ling
    Li, Yu-Kun
    Yang, Hao
    Liu, Xu-Ying
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (06) : 4459 - 4471
  • [34] Robust multi-label feature selection with shared label enhancement
    Li, Yonghao
    Hu, Juncheng
    Gao, Wanfu
    KNOWLEDGE AND INFORMATION SYSTEMS, 2022, 64 (12) : 3343 - 3372
  • [35] Multi-label feature selection based on correlation label enhancement
    He, Zhuoxin
    Lin, Yaojin
    Wang, Chenxi
    Guo, Lei
    Ding, Weiping
    INFORMATION SCIENCES, 2023, 647
  • [36] Multi-label Learning based on Label Entropy Guided Clustering
    Zhang, Ju-Jie
    Fang, Min
    Li, Xiao
    2014 IEEE INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION TECHNOLOGY (CIT), 2014, : 756 - 760
  • [37] Towards Multi-label Feature Selection by Instance and Label Selections
    Mansouri, Dou El Kefel
    Benabdeslem, Khalid
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2021, PT II, 2021, 12713 : 233 - 244
  • [38] Imbalance multi-label data learning with label specific features
    Rastogi, Reshma
    Mortaza, Sayed
    NEUROCOMPUTING, 2022, 513 : 395 - 408
  • [39] Leveraging Supervised Label Dependency Propagation for Multi-label Learning
    Fu, Bin
    Xu, Guandong
    Wang, Zhihai
    Cao, Longbing
    2013 IEEE 13TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2013, : 1061 - 1066
  • [40] LIFT: Multi-Label Learning with Label-Specific Features
    Zhang, Min-Ling
    Wu, Lei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2015, 37 (01) : 107 - 120