Dynamic affinity-based classification of multi-class imbalanced data with one-versus-one decomposition: a fuzzy rough set approach

被引:0
作者
Sarah Vluymans
Alberto Fernández
Yvan Saeys
Chris Cornelis
Francisco Herrera
机构
[1] Ghent University,Department of Applied Mathematics, Computer Science and Statistics
[2] Data Mining and Modeling for Biomedicine,Department of Computer Science and Artificial Intelligence
[3] VIB Inflammation Research Center,Faculty of Computing and Information Technology
[4] University of Granada,undefined
[5] King Abdulaziz University,undefined
来源
Knowledge and Information Systems | 2018年 / 56卷
关键词
Imbalanced data; Multi-class classification; One-versus-one; Fuzzy rough set theory;
D O I
暂无
中图分类号
学科分类号
摘要
Class imbalance occurs when data elements are unevenly distributed among classes, which poses a challenge for classifiers. The core focus of the research community has been on binary-class imbalance, although there is a recent trend toward the general case of multi-class imbalanced data. The IFROWANN method, a classifier based on fuzzy rough set theory, stands out for its performance in two-class imbalanced problems. In this paper, we consider its extension to multi-class data by combining it with one-versus-one decomposition. The latter transforms a multi-class problem into two-class sub-problems. Binary classifiers are applied to these sub-problems, after which their outcomes are aggregated into one prediction. We enhance the integration of IFROWANN in the decomposition scheme in two steps. Firstly, we propose an adaptive weight setting for the binary classifier, addressing the varying characteristics of the sub-problems. We call this modified classifier IFROWANN-WIR\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${{\mathcal {W}}_{\mathrm{IR}}}$$\end{document}. Second, we develop a new dynamic aggregation method called WV–FROST that combines the predictions of the binary classifiers with the global class affinity before making a final decision. In a meticulous experimental study, we show that our complete proposal outperforms the state-of-the-art on a wide range of multi-class imbalanced datasets.
引用
收藏
页码:55 / 84
页数:29
相关论文
共 200 条
[1]  
Abdi L(2016)To combat multi-class imbalanced problems by means of over-sampling techniques IEEE Trans Knowl Data Eng 28 238-251
[2]  
Hashemi S(2015)A proposal for evolutionary fuzzy systems using feature weighting: dealing with overlapping in imbalanced datasets Knowl Based Syst 73 1-17
[3]  
Alshomrani S(2003)Strategies for learning in class imbalance problems Pattern Recog 36 849-851
[4]  
Bawakid A(2004)A study of the behaviour of several methods for balancing machine learning training data SIGKDD Explor 6 20-29
[5]  
Shim S(2014)Dynamic selection of classifiers—a comprehensive review Pattern Recog 47 3665-3680
[6]  
Fernández A(2002)SMOTE: synthetic minority over-sampling technique J Artif Intell Res 16 321-357
[7]  
Herrera F(2016)An empirical study of a hybrid imbalanced-class DT–RST classification procedure to elucidate therapeutic effects in uremia patients Med Biol Eng Comput 54 983-1001
[8]  
Barandela R(2015)A comprehensive study of implicator–conjunctor-based and noise-tolerant fuzzy rough sets: definitions, properties and robustness analysis Fuzzy Sets Syst 275 1-38
[9]  
Sánchez J(2006)Statistical comparisons of classifiers over multiple data sets J Mach Learn Res 7 1-30
[10]  
García V(1990)Rough fuzzy sets and fuzzy rough sets Int J Gen Syst 17 191-209