Re-sampling of multi-class imbalanced data using belief function theory and ensemble learning

被引:15
作者
Grina, Fares [1 ,2 ]
Elouedi, Zied [1 ]
Lefevre, Eric [2 ]
机构
[1] LARODEC, Inst Super Gest Tunis, Tunis, Tunisia
[2] Univ Artois, Lab Genie Informat & Automat Artois LGI2A, UR 3926, F-62400 Bethune, France
关键词
Imbalanced classification; Ensemble learning; Re-sampling; Evidence theory; CLASSIFICATION; SMOTE; PREDICTION;
D O I
10.1016/j.ijar.2023.02.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imbalanced classification refers to problems in which there are significantly more instances available for some classes than for others. Such scenarios require special attention because traditional classifiers tend to be biased towards the majority class which has a large number of examples. Different strategies, such as re-sampling, have been suggested to improve imbalanced learning. Ensemble methods have also been proven to yield promising results in the presence of class-imbalance. However, most of them only deal with binary imbalanced datasets. In this paper, we propose a re-sampling approach based on belief function theory and ensemble learning for dealing with class imbalance in the multi-class setting. This technique assigns soft evidential labels to each instance. This evidential modeling provides more information about each object's region, which improves the selection of objects in both undersampling and oversampling. Our approach firstly selects ambiguous majority instances for undersampling, then oversamples minority objects through the generation of synthetic examples in borderline regions to better improve minority class borders. Finally, to improve the induced results, the proposed re-sampling approach is incorporated into an evidential classifier-independent fusion-based ensemble. The comparative study against well-known ensemble methods reveals that our method is efficient according to the G-Mean and F1-score measures, independently from the chosen classifier. (c) 2023 Elsevier Inc. All rights reserved.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 59 条
[31]  
Li X., 2021, Adv. Neural Inf. Process. Syst., V34
[32]   Exploratory Undersampling for Class-Imbalance Learning [J].
Liu, Xu-Ying ;
Wu, Jianxin ;
Zhou, Zhi-Hua .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2009, 39 (02) :539-550
[33]   Pick and Choose: A GNN-based Imbalanced Learning Approach for Fraud Detection [J].
Liu, Yang ;
Ao, Xiang ;
Qin, Zidi ;
Chi, Jianfeng ;
Feng, Jinghua ;
Yang, Hao ;
He, Qing .
PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2021 (WWW 2021), 2021, :3168-3177
[34]   Credal classification rule for uncertain data based on belief functions [J].
Liu, Zhun-ga ;
Pan, Quan ;
Dezert, Jean ;
Mercier, Gregoire .
PATTERN RECOGNITION, 2014, 47 (07) :2532-2541
[35]   CURE-SMOTE algorithm and hybrid algorithm for feature selection and parameter optimization based on random forests [J].
Ma, Li ;
Fan, Suohai .
BMC BIOINFORMATICS, 2017, 18
[36]  
Mahalanobis P.C., 1936, P NATL I SCI INDIA, V2, P49, DOI [10.1007/s13171-019-00164-5, DOI 10.1007/S13171-019-00164-5]
[37]  
More AS, 2017, 2017 1ST INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND INFORMATION MANAGEMENT (ICISIM), P72, DOI 10.1109/ICISIM.2017.8122151
[38]   OAHO: an effective algorithm for multi-class learning from imbalanced data [J].
Murphey, Yi L. ;
Wang, Haoxing ;
Ou, Guobin ;
Feldkamp, Lee A. .
2007 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-6, 2007, :406-+
[39]   Imbalance Data Classification Based on Belief Function Theory [J].
Niu, Jiawei ;
Liu, Zhunga .
BELIEF FUNCTIONS: THEORY AND APPLICATIONS (BELIEF 2021), 2021, 12915 :96-104
[40]   Fast-CBUS: A fast clustering-based undersampling method for addressing the class imbalance problem [J].
Ofek, Nir ;
Rokach, Lior ;
Stern, Roni ;
Shabtai, Asaf .
NEUROCOMPUTING, 2017, 243 :88-102