Self-adaptive kernel K-means algorithm based on the shuffled frog leaping algorithm

被引:12
作者
Fan, Shuyan [1 ]
Ding, Shifei [1 ,2 ]
Xue, Yu [3 ]
机构
[1] China Univ Min & Technol, Sch Comp Sci & Technol, Xuzhou 221116, Peoples R China
[2] Chinese Acad Sci, Inst Comp Technol, Key Lab Intelligent Informat Proc, Beijing 100090, Peoples R China
[3] Nanjing Univ Informat Sci & Technol, Sch Comp & Software, Nanjing 210044, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Kernel K-means; Shuffled frog leaping algorithm; Clustering validity index; Clustering analysis; OPTIMIZATION; EVOLUTION;
D O I
10.1007/s00500-016-2389-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Kernel K-means can handle nonlinearly separate datasets by mapping the input datasets into a high-dimensional feature space. The kernel matrix reflects the inner structure of data, so it is a key to construct an appropriate kernel matrix. However, many kernel-based methods need to be set kernel parameter artificially in advance. It is difficult to set an appropriate kernel parameter for each dataset artificially, which limits the performance of the kernel K-means algorithm to some extent. It is necessary to design a method which can adjust the kernel parameter automatically according to the data structure. In addition, the number of clusters also needs to be set. To overcome these challenges, this paper proposed a self-adaptive kernel K-means based on the shuffled frog leaping algorithm, which regard the kernel parameter and the number of clusters as the position information of the frog. We designed a clustering validity index named Between-Within Proportion suitable for the kernel space (KBWP) by modifying the clustering validity index Between-Within Proportion (BWP). Treat KBWP as fitness in the shuffled frog leaping algorithm, and then do local and global optimization until the max iterations. The kernel parameter and the number of clusters corresponding to the maximum fitness are optimal. We experimentally verify our algorithm on artificial datasets and real datasets. Experimental results demonstrate the effectiveness and good performance of the proposed algorithm.
引用
收藏
页码:861 / 872
页数:12
相关论文
共 33 条
[1]  
[Anonymous], 2005, Advances in Neural Information Processing Systems
[2]  
Bach FR, 2004, ADV NEUR IN, V16, P305
[3]  
Back T., 1996, EVOLUTIONARY ALGORIT, DOI DOI 10.1093/OSO/9780195099713.001.0001
[4]   Genetic clustering for automatic evolution of clusters and application to image classification [J].
Bandyopadhyay, S ;
Maulik, U .
PATTERN RECOGNITION, 2002, 35 (06) :1197-1208
[5]   Automatic kernel clustering with a Multi-Elitist Particle Swarm Optimization Algorithm [J].
Das, Swagatam ;
Abraham, Ajith ;
Konar, Amit .
PATTERN RECOGNITION LETTERS, 2008, 29 (05) :688-699
[6]   Optimization of water distribution network design using the Shuffled Frog Leaping Algorithm [J].
Eusuff, MM ;
Lansey, KE .
JOURNAL OF WATER RESOURCES PLANNING AND MANAGEMENT, 2003, 129 (03) :210-225
[7]   A survey of kernel and spectral methods for clustering [J].
Filippone, Maurizio ;
Camastra, Francesco ;
Masulli, Francesco ;
Rovetta, Stefano .
PATTERN RECOGNITION, 2008, 41 (01) :176-190
[8]   Clustering by passing messages between data points [J].
Frey, Brendan J. ;
Dueck, Delbert .
SCIENCE, 2007, 315 (5814) :972-976
[9]   Structural Minimax Probability Machine [J].
Gu, Bin ;
Sun, Xingming ;
Sheng, Victor S. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (07) :1646-1656
[10]   Incremental Support Vector Learning for Ordinal Regression [J].
Gu, Bin ;
Sheng, Victor S. ;
Tay, Keng Yeow ;
Romano, Walter ;
Li, Shuo .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2015, 26 (07) :1403-1416