A fast online learning algorithm of radial basis function network with locality sensitive hashing

被引:4
作者
Ali S.H.A. [1 ,2 ]
Fukase K. [1 ]
Ozawa S. [1 ]
机构
[1] Graduate School of Engineering, Kobe University, 1-1 Rokko-Dai, Kobe
[2] Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia (UTHM), Parit Raja, Batu Pahat, 86400, Johor
基金
日本学术振兴会;
关键词
Incremental learning; Large-scale data sequence; Locality sensitive hashing; Resource allocating network;
D O I
10.1007/s12530-015-9141-5
中图分类号
学科分类号
摘要
In this paper, we propose a new incremental learning algorithm of radial basis function (RBF) Network to accelerate the learning for large-scale data sequence. Along with the development of the internet and sensor technologies, a time series of large data chunk are continuously generated in our daily life. Thus it is usually difficult to learn all the data within a short period. A remedy for this is to select only essential data from a given data chunk and provide them to a classifier model to learn. In the proposed method, only data in untrained regions, which correspond to a region with a low output margin, are selected. The regions are formed by grouping the data based on their near neighbor using locality sensitive hashing (LSH), in which LSH has been developed to search neighbors quickly in an approximated way. As the proposed method does not use all training data to calculate the output margins, the time of the data selection is expected to be shortened. In the incremental learning phase, in order to suppress catastrophic forgetting, we also exploit LSH to select neighbor RBF units quickly. In addition, we propose a method to update the hash table in LSH so that the data selection can be adaptive during the learning. From the performance of nine datasets, we confirm that the proposed method can learn large-scale data sequences fast without sacrificing the classification accuracies. This fact implies that the data selection and the incremental learning work effectively in the proposed method. © 2016, Springer-Verlag Berlin Heidelberg.
引用
收藏
页码:173 / 186
页数:13
相关论文
共 26 条
[1]  
Andoni A., Indyk P., Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions, Commun ACM, 51, 1, pp. 117-122, (2008)
[2]  
Attar V., Sinha P., Wankhade K., A fast and light classifier for data streams, Evol Syst, 1, 3, pp. 199-207, (2010)
[3]  
Bache K., Lichman M., UCI Machine Learning Repository, (2013)
[4]  
Casey M., Slaney M., Fast recognition of remixed music audio. In Proceedings, IEEE ICASSP. IV, pp. 1425-1428, (2007)
[5]  
Cauwenberghs G., Poggio T., Incremental and decremental support vector machine learning, Adv Neural Inf Process Syst (NIPS’2000), (2000)
[6]  
Chellapilla K., Mityagin A., Charles D.X., GigaHash: scalable minimal perfect hashing for billions of urls, In: Proceedings of the 16th international conference on World Wide Web, (2007)
[7]  
Dan R., Kevin S., Margin-based active learning for structured output spaces. In: Proceedings ECML’06, p 413–424, (2006)
[8]  
Datar M., Immorlica N., Indyk P., Mirrokni V.S., Locality-sensitive hashing scheme based on p-stable distributions. In: Proceedings of symposium on computational geometry (SoCG’04), p 253–262, (2004)
[9]  
Davies D.L., Bouldin D.W., A cluster separation measure, IEEE Trans Pattern Anal Mach Intell, 1, 2, pp. 224-227, (1979)
[10]  
Diehl C.P., Cauwenberghs G., SVM incremental learning, adaptation and optimization. In: Proceedings of the 2003 international joint conference on neural networks (IJCNN’03), p 2685–2690, (2003)