Learning Imbalanced Classifiers Locally and Globally with One-Side Probability Machine

被引:0
作者
Kaizhu Huang
Rui Zhang
Xu-Cheng Yin
机构
[1] Xi’an Jiaotong-Liverpool University,Department of EEE
[2] SIP,Mathematics and Physics Center
[3] Xi’an Jiaotong-Liverpool University,Department of Computer Science and Technology, School of Computer and Communication Engineering
[4] SIP,undefined
[5] University of Science and Technology Beijing,undefined
来源
Neural Processing Letters | 2015年 / 41卷
关键词
Learning locally and globally; Imbalanced learning ; Classification;
D O I
暂无
中图分类号
学科分类号
摘要
We consider the imbalanced learning problem, where the data associated with one class are far fewer than those associated with the other class. Current imbalanced learning methods often handle this problem by adapting certain intermediate parameters so as to impose a bias on the minority data. However, most of these methods are in rigorous and need to adapt those factors via the trial-and-error procedure. Recently, a new model called Biased Minimax Probability Machine (BMPM) presents a rigorous and systematic work and has demonstrated very promising performance on imbalance learning. Despite its success, BMPM exclusively relies on global information, namely, the first order and second order data information; such information might be however unreliable, especially for the minority data. In this paper, we propose a new model called One-Side Probability Machine (OSPM). Different from the previous approaches, OSPM can lead to rigorous treatment on biased classification tasks. Importantly, the proposed OSPM exploits the reliable global information from one side only, i.e., the majority class, while engaging the robust local learning from the other side, i.e., the minority class. To our best knowledge, OSPM presents the first model capable of learning data both locally and globally. Our proposed model has also established close connections with various famous models such as BMPM, Support Vector Machine, and Maxi-Min Margin Machine. One appealing feature is that the optimization problem involved in the novel OSPM model can be cast as a convex second order conic programming problem with the global optimum guaranteed. A series of experimental results on three data sets demonstrate the advantages of our proposed methods over four competitive approaches.
引用
收藏
页码:311 / 323
页数:12
相关论文
共 45 条
[1]  
Chawla N(2002)Smote: sythetic minority over-sampling technique J Artif Intell Res 16 321-357
[2]  
Bowyer K(2009)Learning from imbalanced data IEEE Trans Knowl Data Eng 21 1263-1284
[3]  
Hall L(2006)Maximizing sensitivity in medical diagnosis using biased minimax probability machine IEEE Trans Biomed Eng 53 821-831
[4]  
Kegelmeyer W(2008)Maxi-min margin machine: learning large margin classifiers globally and locally IEEE Trans Neural Netw 19 260-272
[5]  
He H(2006)Imbalanced learning with biased minimax probability machine IEEE Trans Syst Man Cybern Part B 36 913-923
[6]  
Garcia E(2004)The minimum error minimax probability machine J Mach Learn Res 5 1253-1286
[7]  
Huang K(2010)Sparse learning for support vector classification Pattern Recognit Lett 31 1944-1951
[8]  
Yang H(2014)Cost-sensitive decision tree ensembles for effective imbalanced classification Appl Soft Comput 14 554-562
[9]  
King I(2002)A robust minimax approach to classification J Mach Learn Res 3 555-582
[10]  
Lyu MR(1998)Applications of second order cone programming Linear Algebra Appl 284 193-228