Differential evolution based nearest prototype classifier with optimized distance measures for the features in the data sets

被引:18
作者
Koloseni, David [1 ,4 ]
Lampinen, Jouni [2 ,3 ]
Luukka, Pasi [1 ,5 ]
机构
[1] Lappeenranta Univ Technol, Lab Appl Math, FI-53851 Lappeenranta, Finland
[2] Univ Vaasa, Dept Comp Sci, FI-65101 Vaasa, Finland
[3] VSB Tech Univ Ostrava, Dept Comp Sci, Ostrava 70833, Czech Republic
[4] Univ Dar Es Salaam, Dept Math, Dar Es Salaam, Tanzania
[5] Lappeenranta Univ Technol, Sch Business, FI-53851 Lappeenranta, Finland
关键词
Differential evolution; Classification; Distance measures; Distance selection for the feature; Pool of distances; NEURAL-NETWORKS; ALGORITHM;
D O I
10.1016/j.eswa.2013.01.040
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper a further generalization of differential evolution based data classification method is proposed, demonstrated and initially evaluated. The differential evolution classifier is a nearest prototype vector based classifier that applies a global optimization algorithm, differential evolution, for determining the optimal values for all free parameters of the classifier model during the training phase of the classifier. The earlier version of differential evolution classifier that applied individually optimized distance measure for each new data set to be classified is generalized here so, that instead of optimizing a single distance measure for the given data set, we take a further step by proposing an approach where distance measures are optimized individually for each feature of the data set to be classified. In particular, distance measures for each feature are selected optimally from a predefined pool of alternative distance measures. The optimal distance measures are determined by differential evolution algorithm, which is also determining the optimal values for all free parameters of the selected distance measures in parallel. After determining the optimal distance measures for each feature together with their optimal parameters, we combine all featurewisely determined distance measures to form a single total distance measure, that is to be applied for the final classification decisions. The actual classification process is still based on the nearest prototype vector principle; A sample belongs to the class represented by the nearest prototype vector when measured with the above referred optimized total distance measure. During the training process the differential evolution algorithm determines optimally the class vectors, selects optimal distance metrics for each data feature, and determines the optimal values for the free parameters of each selected distance measure. Based on experimental results with nine well known classification benchmark data sets, the proposed approach yield a statistically significant improvement to the classification accuracy of differential evolution classifier. (C) 2013 Elsevier Ltd. All rights reserved.
引用
收藏
页码:4075 / 4082
页数:8
相关论文
共 40 条
[31]  
Price Kenneth V., 1999, NEW IDEAS OPTIMIZATI, P79, DOI 10.5555/329055.329069
[32]  
Schonfeld J, 2006, IEEE C EVOL COMPUTAT, P2316
[33]   Comparison of distance measures in spatial analytical modeling for health service planning [J].
Shahid, Rizwan ;
Bertazzon, Stefania ;
Knudtson, Merril L. ;
Ghali, William A. .
BMC HEALTH SERVICES RESEARCH, 2009, 9
[34]   Differential evolution - A simple and efficient heuristic for global optimization over continuous spaces [J].
Storn, R ;
Price, K .
JOURNAL OF GLOBAL OPTIMIZATION, 1997, 11 (04) :341-359
[35]   Classification rule discovery with DE/QDE algorithm [J].
Su, Haijun ;
Yang, Yupu ;
Zhao, Liang .
EXPERT SYSTEMS WITH APPLICATIONS, 2010, 37 (02) :1216-1222
[36]   DE/EDA: A new evolutionary algorithm for global optimization [J].
Sun, JY ;
Zhang, QF ;
Tsang, EPK .
INFORMATION SCIENCES, 2005, 169 (3-4) :249-262
[37]   Differential evolution for optimizing the positioning of prototypes in nearest neighbor classification [J].
Triguero, Isaac ;
Garcia, Salvador ;
Herrera, Francisco .
PATTERN RECOGNITION, 2011, 44 (04) :901-916
[38]  
van der Heijden F., 2004, CLASSIFICATION PARAM
[39]  
Yu JW, 2007, ICNC 2007: THIRD INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION, VOL 3, PROCEEDINGS, P456
[40]  
Zmudaa M. A., 2009, APPL SOFT COMPUT, V2, P269