Instance reduction for one-class classification

被引:24
作者
Krawczyk, Bartosz [1 ]
Triguero, Isaac [2 ]
Garcia, Salvador [3 ]
Wozniak, Michal [4 ]
Herrera, Francisco [3 ,5 ]
机构
[1] Virginia Commonwealth Univ, Dept Comp Sci, Richmond, VA 23284 USA
[2] Univ Nottingham, Sch Comp Sci, Automated Scheduling Optimisat & Planning ASAP Gr, Nottingham, England
[3] Univ Granada, Dept Comp Sci & Artificial Intelligence, CITIC UGR, Granada, Spain
[4] Wroclaw Univ Technol, Dept Syst & Comp Networks, Comp Sci, Wroclaw, Poland
[5] King Abdulaziz Univ, Fac Comp & Informat Technol, Jeddah, Saudi Arabia
关键词
Machine learning; One-class classification; Instance reduction; Training set selection; Evolutionary computing; DIFFERENTIAL EVOLUTION; SELECTION;
D O I
10.1007/s10115-018-1220-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Instance reduction techniques are data preprocessing methods originally developed to enhance the nearest neighbor rule for standard classification. They reduce the training data by selecting or generating representative examples of a given problem. These algorithms have been designed and widely analyzed in multi-class problems providing very competitive results. However, this issue was rarely addressed in the context of one-class classification. In this specific domain a reduction of the training set may not only decrease the classification time and classifier's complexity, but also allows us to handle internal noisy data and simplify the data description boundary. We propose two methods for achieving this goal. The first one is a flexible framework that adjusts any instance reduction method to one-class scenario by introduction of meaningful artificial outliers. The second one is a novel modification of evolutionary instance reduction technique that is based on differential evolution and uses consistency measure for model evaluation in filter or wrapper modes. It is a powerful native one-class solution that does not require an access to counterexamples. Both of the proposed algorithms can be applied to any type of one-class classifier. On the basis of extensive computational experiments, we show that the proposed methods are highly efficient techniques to reduce the complexity and improve the classification performance in one-class scenarios.
引用
收藏
页码:601 / 628
页数:28
相关论文
共 59 条
[1]  
AHA DW, 1991, MACH LEARN, V6, P37, DOI 10.1007/BF00153759
[2]   Prototype-Based Domain Description for One-Class Classification [J].
Angiulli, Fabrizio .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (06) :1131-1144
[3]  
[Anonymous], 2009, Memetic Comp, DOI DOI 10.1007/S12293-009-0008-9
[4]   Soft clustering using weighted one-class support vector machines [J].
Bicego, Manuele ;
Figueiredo, Mario A. T. .
PATTERN RECOGNITION, 2009, 42 (01) :27-32
[5]  
Bolon-Canedo V., 2015, ARTIFICIAL INTELLIGE, V1, P1
[6]   Self-adapting control parameters in differential evolution: A comparative study on numerical benchmark problems [J].
Brest, Janez ;
Greiner, Saso ;
Boskovic, Borko ;
Mernik, Marjan ;
Zumer, Vijern .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2006, 10 (06) :646-657
[7]  
Cabral GG, 2011, IEEE SYS MAN CYBERN, P983, DOI 10.1109/ICSMC.2011.6083797
[8]   Multi-objective genetic programming for feature extraction and data visualization [J].
Cano, Alberto ;
Ventura, Sebastian ;
Cios, Krzysztof J. .
SOFT COMPUTING, 2017, 21 (08) :2069-2089
[9]   Using evolutionary algorithms as instance selection for data reduction in KDD: An experimental study [J].
Cano, JR ;
Herrera, F ;
Lozano, M .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2003, 7 (06) :561-575
[10]  
Chen YH, 2009, J MACH LEARN RES, V10, P747