Selecting Critical Patterns Based on Local Geometrical and Statistical Information

被引:117
作者
Li, Yuhua [1 ]
Maguire, Liam [1 ]
机构
[1] Univ Ulster, Sch Comp & Intelligent Syst, Magee BT48 7LJ, Londonderry, North Ireland
关键词
Pattern selection; data reduction; border pattern; edge pattern; NEAREST-NEIGHBOR CLASSIFICATION; PROTOTYPE SELECTION; SEARCH ALGORITHM; CLASSIFIERS; REDUCTION; SUBSET; OPTIMIZATION; NETWORKS; CHOICE;
D O I
10.1109/TPAMI.2010.188
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pattern selection methods have been traditionally developed with a dependency on a specific classifier. In contrast, this paper presents a method that selects critical patterns deemed to carry essential information applicable to train those types of classifiers which require spatial information of the training data set. Critical patterns include those edge patterns that define the boundary and those border patterns that separate classes. The proposed method selects patterns from a new perspective, primarily based on their location in input space. It determines class edge patterns with the assistance of the approximated tangent hyperplane of a class surface. It also identifies border patterns between classes using local probability. The proposed method is evaluated on benchmark problems using popular classifiers, including multilayer perceptrons, radial basis functions, support vector machines, and nearest neighbors. The proposed approach is also compared with four state-of-the-art approaches and it is shown to provide similar but more consistent accuracy from a reduced data set. Experimental results demonstrate that it selects patterns sufficient to represent class boundary and to preserve the decision surface.
引用
收藏
页码:1189 / 1201
页数:13
相关论文
共 69 条
[1]   Fast multidimensional nearest neighbor search algorithm using priority queue [J].
Ajoka, Shiro ;
Tsuge, Satoru ;
Shishibori, Masami ;
Kita, Kenji .
ELECTRICAL ENGINEERING IN JAPAN, 2008, 164 (03) :69-77
[2]   Fast nearest neighbor condensation for large data sets classification [J].
Angiulli, Fabrizio .
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2007, 19 (11) :1450-1464
[3]   Condensed nearest neighbor data domain description [J].
Angiulli, Fabrizio .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2007, 29 (10) :1746-1758
[4]   Decision boundary preserving prototype selection for nearest neighbor classification [J].
Barandela, R ;
Ferri, FJ ;
Sánchez, JS .
INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2005, 19 (06) :787-806
[5]  
BENTLEY JL, 1979, COMPUT SURV, V11, P397, DOI 10.1145/356789.356797
[6]   Nearest prototype classifier designs: An experimental study [J].
Bezdek, JC ;
Kuncheva, LI .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2001, 16 (12) :1445-1473
[7]  
Blake C. L., 1998, Uci repository of machine learning databases
[8]  
Broomhead D. S., 1988, Complex Systems, V2, P321
[9]   A tutorial on Support Vector Machines for pattern recognition [J].
Burges, CJC .
DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) :121-167
[10]   PEDAGOGICAL PATTERN SELECTION-STRATEGIES [J].
CACHIN, C .
NEURAL NETWORKS, 1994, 7 (01) :175-181