GOBoost: G-mean Optimized Boosting Framework for Class Imbalance Learning

被引:0
作者
Lu, Yang [1 ]
Cheung, Yiu-ming [1 ,2 ]
Tang, Yuan Yan [3 ]
机构
[1] Hong Kong Baptist Univ, Dept Comp Sci, Hong Kong, Hong Kong, Peoples R China
[2] HKBU Inst Res & Continuing Educ, Shenzhen, Peoples R China
[3] Univ Macau, Fac Sci & Technol, Dept Comp & Informat Sci, Macau, Peoples R China
来源
PROCEEDINGS OF THE 2016 12TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION (WCICA) | 2016年
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Boosting-based methods are effective for class imbalance problem, where the numbers of samples in two or more classes are severely unequal. However, the classifier weights of existing boosting-based methods are calculated by minimizing the error rate, which is inconsistent with the objective of class imbalance learning. As a result, the classifier weights cannot represent the performance of individual classifiers properly when the data is imbalanced. In this paper, we therefore propose a G-mean Optimized Boosting (GOBoost) framework to assign classifier weights optimized on G-mean. Subsequently, high weights are assigned to the classifier with high accuracy on both the majority class and the minority class. The GOBoost framework can be applied to any AdaBoost-based method for class imbalance learning by simply replacing the calculation of classifier weights. Accordingly, we extend six AdaBoost-based methods to GOBoost-based methods for comparative studies in class imbalance learning. The experiments conducted on 12 real class imbalance data sets show that GOBoost-based methods significantly outperform the corresponding AdaBoost-based methods in terms of F1 and G-mean metrics.
引用
收藏
页码:3149 / 3154
页数:6
相关论文
共 18 条
[1]  
Breiman F, 1984, OLSHEN STONE CLASSIF
[2]   SMOTE: Synthetic minority over-sampling technique [J].
Chawla, Nitesh V. ;
Bowyer, Kevin W. ;
Hall, Lawrence O. ;
Kegelmeyer, W. Philip .
2002, American Association for Artificial Intelligence (16)
[3]   SMOTEBoost: Improving prediction of the minority class in boosting [J].
Chawla, NV ;
Lazarevic, A ;
Hall, LO ;
Bowyer, KW .
KNOWLEDGE DISCOVERY IN DATABASES: PKDD 2003, PROCEEDINGS, 2003, 2838 :107-119
[4]  
Demsar J, 2006, J MACH LEARN RES, V7, P1
[5]  
Fan W, 1999, MACHINE LEARNING, PROCEEDINGS, P97
[6]   A decision-theoretic generalization of on-line learning and an application to boosting [J].
Freund, Y ;
Schapire, RE .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) :119-139
[7]   A Review on Ensembles for the Class Imbalance Problem: Bagging-, Boosting-, and Hybrid-Based Approaches [J].
Galar, Mikel ;
Fernandez, Alberto ;
Barrenechea, Edurne ;
Bustince, Humberto ;
Herrera, Francisco .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2012, 42 (04) :463-484
[8]   EUSBoost: Enhancing ensembles for highly imbalanced data-sets by evolutionary undersampling [J].
Galar, Mikel ;
Fernandez, Alberto ;
Barrenechea, Edurne ;
Herrera, Francisco .
PATTERN RECOGNITION, 2013, 46 (12) :3460-3471
[9]   Learning from Imbalanced Data [J].
He, Haibo ;
Garcia, Edwardo A. .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2009, 21 (09) :1263-1284
[10]  
Kubat M, 1997, LECT NOTES ARTIF INT, V1224, P146