A simple plug-in bagging ensemble based on threshold-moving for classifying binary and multiclass imbalanced data

被引:111
作者
Collell, Guillem [1 ,4 ]
Prelec, Drazen [1 ,2 ,3 ]
Patil, Kaustubh R. [1 ,5 ]
机构
[1] MIT, MIT Sloan Neuroecon Lab, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[2] MIT, Dept Econ, Cambridge, MA 02139 USA
[3] MIT, Brain & Cognit Sci, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[4] Katholieke Univ Leuven, Dept Comp Sci, B-3001 Heverlee, Belgium
[5] Res Ctr Julich, Inst Neurosci & Med, Brain & Behav INM 7, D-52425 Julich, Germany
基金
英国惠康基金;
关键词
Imbalanced data; Binary classification; Multiclass classification; Bagging ensembles; Resampling; Posterior calibration; CLASSIFIERS;
D O I
10.1016/j.neucom.2017.08.035
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Class imbalance presents a major hurdle in the application of classification methods. A commonly taken approach is to learn ensembles of classifiers using rebalanced data. Examples include bootstrap averaging ( bagging) combined with either undersampling or oversampling of the minority class examples. However, rebalancing methods entail asymmetric changes to the examples of different classes, which in turn can introduce their own biases. Furthermore, these methods often require specifying the performance measure of interest a priori, i.e., before learning. An alternative is to employ the threshold moving technique, which applies a threshold to the continuous output of a model, offering the possibility to adapt to a performance measure a posteriori, i.e., a plug-in method. Surprisingly, little attention has been paid to this combination of a bagging ensemble and threshold-moving. In this paper, we study this combination and demonstrate its competitiveness. Contrary to the other resampling methods, we preserve the natural class distribution of the data resulting in well-calibrated posterior probabilities. Additionally, we extend the proposed method to handle multiclass data. We validated our method on binary and multiclass benchmark data sets by using both, decision trees and neural networks as base classifiers. We perform analyses that provide insights into the proposed method. (C) 2017 The Author(s). Published by Elsevier B.V.
引用
收藏
页码:330 / 340
页数:11
相关论文
共 43 条
[1]  
[Anonymous], 2004, COMBINING PATTERN CL, DOI DOI 10.1002/0471660264
[2]  
[Anonymous], P IEEE S COMP INT DA
[3]  
[Anonymous], 2009, ELEMENTS STAT LEARNI
[4]  
Batista GE., 2004, ACM SIGKDD EXPL NEWS, V6, P20, DOI [DOI 10.1145/1007730.1007735, 10.1145/1007730.1007735]
[5]   An empirical comparison of voting classification algorithms: Bagging, boosting, and variants [J].
Bauer, E ;
Kohavi, R .
MACHINE LEARNING, 1999, 36 (1-2) :105-139
[6]  
Blaszczynski J., 2013, P 8 INT C COMP REC S
[7]   Bagging predictors [J].
Breiman, L .
MACHINE LEARNING, 1996, 24 (02) :123-140
[8]  
Chawla N., 2005, P 1 INT WORKSH UT BA
[9]   SMOTE: Synthetic minority over-sampling technique [J].
Chawla, Nitesh V. ;
Bowyer, Kevin W. ;
Hall, Lawrence O. ;
Kegelmeyer, W. Philip .
2002, American Association for Artificial Intelligence (16)
[10]   Calibrating Probability with Undersampling for Unbalanced Classification [J].
Dal Pozzolo, Andrea ;
Caelen, Olivier ;
Johnson, Reid A. ;
Bontempi, Gianluca .
2015 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2015, :159-166