A novel progressively undersampling method based on the density peaks sequence for imbalanced data

被引:60
作者
Xie, Xiaoying [1 ]
Liu, Huawen [2 ]
Zeng, Shouzhen [3 ]
Lin, Lingbin [4 ]
Li, Wen [5 ]
机构
[1] Zhejiang Normal Univ, Coll Econ & Management, Jinhua 321004, Zhejiang, Peoples R China
[2] Zhejiang Normal Univ, Coll Math & Comp Sci, Jinhua 321004, Zhejiang, Peoples R China
[3] Ningbo Univ, Sch Business, Ningbo 315211, Peoples R China
[4] Zhejiang Normal Univ, Student Management Off, Jinhua 321004, Zhejiang, Peoples R China
[5] Curtin Univ, Dept Math & Stat, Perth, WA 6845, Australia
关键词
Progressive undersampling; Density peaks sequence; Importance degree; Optimal undersampling size; Imbalanced data;
D O I
10.1016/j.knosys.2020.106689
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Undersampling is a widely used resampling technique for imbalanced data. As traditional undersampling techniques, typically making majority and minority classes in imbalanced data into the same scale, tend to miss valuable information, many strategies like clustering have been developed. However, two essential problems still remain and require more efforts to be put; that is, which and how many instances should be extracted in undersampling. To alleviate these two problems, in this paper we propose a novel undersampling method for imbalanced data. It exploits a sequence of density peaks to progressively extract instances from the majority classes of the imbalanced data. Specifically, two factors are introduced to measure the importance degree of each instance in the majority classes. With these two factors, we generate a sampling sequence based on the importance of instances for classification. Furthermore, the optimal undersampling size of the majority classes is automatically determined by progressively extracting the important instances from the sequence. To evaluate the effectiveness of the proposed method, a series of experiments comparing to six popular undersampling methods were conducted on 40 public benchmark datasets. The experimental results show that the performance of the proposed undersampling method is superior to the state-of-the-art undersampling methods. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:11
相关论文
共 47 条
[1]  
Alcalá-Fdez J, 2011, J MULT-VALUED LOG S, V17, P255
[2]  
[Anonymous], 2004, ACM Sigkdd Explorations Newsletter
[3]  
Bobkov Sergey, 2019, ONE DIMENSIONAL EMPI, V261
[4]  
Chen C, 2011, 2011 IEEE INTERNATIONAL CONFERENCE ON INFORMATION REUSE AND INTEGRATION (IRI), P384, DOI 10.1109/IRI.2011.6009578
[5]   Progressive Multi-Jittered Sample Sequences [J].
Christensen, Per ;
Kensler, Andrew ;
Kilpatrick, Charlie .
COMPUTER GRAPHICS FORUM, 2018, 37 (04) :21-33
[6]   Entropy-based fuzzy support vector machine for imbalanced datasets [J].
Fan, Qi ;
Wang, Zhe ;
Li, Dongdong ;
Gao, Daqi ;
Zha, Hongyuan .
KNOWLEDGE-BASED SYSTEMS, 2017, 115 :87-99
[7]  
Fernandez A., 2018, Learning from Imbalanced Data Sets, DOI DOI 10.1007/978-3-319-98074-4
[8]   Analysing the classification of imbalanced data-sets with multiple classes: Binarization techniques and ad-hoc approaches [J].
Fernandez, Alberto ;
Lopez, Victoria ;
Galar, Mikel ;
Jose del Jesus, Maria ;
Herrera, Francisco .
KNOWLEDGE-BASED SYSTEMS, 2013, 42 :97-110
[9]   A comprehensive data level analysis for cancer diagnosis on imbalanced data [J].
Fotouhi, Sara ;
Asadi, Shahrokh ;
Kattan, Michael W. .
JOURNAL OF BIOMEDICAL INFORMATICS, 2019, 90
[10]   A survey on instance selection for active learning [J].
Fu, Yifan ;
Zhu, Xingquan ;
Li, Bin .
KNOWLEDGE AND INFORMATION SYSTEMS, 2013, 35 (02) :249-283