A fuzzy K-nearest neighbor classifier to deal with imperfect data

被引:17
|
作者
Cadenas, Jose M. [1 ]
Carmen Garrido, M. [1 ]
Martinez, Raquel [2 ]
Munoz, Enrique [3 ]
Bonissone, Piero P. [4 ]
机构
[1] Univ Murcia, Dept Informat & Commun Engn, Murcia, Spain
[2] Catholic Univ Murcia, Dept Comp Engn, Murcia, Spain
[3] Univ Milan, Dept Comp Sci, Crema, Italy
[4] Piero P Bonissone Analyt LLC, San Diego, CA USA
关键词
k-nearest neighbors; Classification; Imperfect data; Distance/dissimilarity measures; Combination methods; PERFORMANCE; RULES; ALGORITHMS;
D O I
10.1007/s00500-017-2567-x
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The k-nearest neighbors method (kNN) is a nonparametric, instance-based method used for regression and classification. To classify a new instance, the kNN method computes its k nearest neighbors and generates a class value from them. Usually, this method requires that the information available in the datasets be precise and accurate, except for the existence of missing values. However, data imperfection is inevitable when dealing with real-world scenarios. In this paper, we present the kNN(imp) classifier, a k-nearest neighbors method to perform classification from datasets with imperfect value. The importance of each neighbor in the output decision is based on relative distance and its degree of imperfection. Furthermore, by using external parameters, the classifier enables us to define the maximum allowed imperfection, and to decide if the final output could be derived solely from the greatest weight class (the best class) or from the best class and a weighted combination of the closest classes to the best one. To test the proposed method, we performed several experiments with both synthetic and real-world datasets with imperfect data. The results, validated through statistical tests, show that the kNN(imp) classifier is robust when working with imperfect data and maintains a good performance when compared with other methods in the literature, applied to datasets with or without imperfection.
引用
收藏
页码:3313 / 3330
页数:18
相关论文
共 50 条
  • [41] Leukocyte Classification based on Transfer Learning of VGG16 Features by K-Nearest Neighbor Classifier
    Baby, Diana
    Devaraj, Sujitha Juliet
    Raj M. M., Anishin
    ICSPC'21: 2021 3RD INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION (ICPSC), 2021, : 252 - 256
  • [42] Classification of Power Quality Disturbances Using Wavelet Transform and K-Nearest Neighbor Classifier
    Ngo Minh Khoa
    Dinh Thanh Viet
    Nguyen Huu Hieu
    2013 IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS (ISIE), 2013,
  • [43] Approximate direct and reverse nearest neighbor queries, and the k-nearest neighbor graph
    Figueroa, Karina
    Paredes, Rodrigo
    SISAP 2009: 2009 SECOND INTERNATIONAL WORKSHOP ON SIMILARITY SEARCH AND APPLICATIONS, PROCEEDINGS, 2009, : 91 - +
  • [44] Performance study of K-nearest neighbor classifier and K-means clustering for predicting the diagnostic accuracy
    Mittal K.
    Aggarwal G.
    Mahajan P.
    International Journal of Information Technology, 2019, 11 (3) : 535 - 540
  • [45] Exact bootstrap k-nearest neighbor learners
    Steele, Brian M.
    MACHINE LEARNING, 2009, 74 (03) : 235 - 255
  • [46] Evidential instance selection for K-nearest neighbor classification of big data
    Gong, Chaoyu
    Su, Zhi-gang
    Wang, Pei-hong
    Wang, Qian
    You, Yang
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2021, 138 : 123 - 144
  • [47] EVOLVING EDITED k-NEAREST NEIGHBOR CLASSIFIERS
    Gil-Pita, Roberto
    Yao, Xin
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2008, 18 (06) : 459 - 467
  • [48] Asymptotics of k-nearest Neighbor Riesz Energies
    Hardin, Douglas P.
    Saff, Edward B.
    Vlasiuk, Oleksandr
    CONSTRUCTIVE APPROXIMATION, 2024, 59 (02) : 333 - 383
  • [49] Exact bootstrap k-nearest neighbor learners
    Brian M. Steele
    Machine Learning, 2009, 74 : 235 - 255
  • [50] Kinetic Reverse k-Nearest Neighbor Problem
    Rahmati, Zahed
    King, Valerie
    Whitesides, Sue
    COMBINATORIAL ALGORITHMS, IWOCA 2014, 2015, 8986 : 307 - 317