Generating feature spaces for linear algorithms with regularized sparse kernel slow feature analysis

被引:16
作者
Boehmer, Wendelin [1 ]
Gruenewaelder, Steffen [2 ]
Nickisch, Hannes [3 ]
Obermayer, Klaus [1 ]
机构
[1] Tech Univ Berlin, Neural Informat Proc Grp, Berlin, Germany
[2] UCL, Ctr Computat Stat & Machine Learning, London, England
[3] Philips Res Labs, Hamburg, Germany
基金
英国工程与自然科学研究理事会;
关键词
Time series; Latent variables; Unsupervised learning; Slow feature analysis; Sparse kernel methods; Linear classification;
D O I
10.1007/s10994-012-5300-0
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Without non-linear basis functions many problems can not be solved by linear algorithms. This article proposes a method to automatically construct such basis functions with slow feature analysis (SFA). Non-linear optimization of this unsupervised learning method generates an orthogonal basis on the unknown latent space for a given time series. In contrast to methods like PCA, SFA is thus well suited for techniques that make direct use of the latent space. Real-world time series can be complex, and current SFA algorithms are either not powerful enough or tend to over-fit. We make use of the kernel trick in combination with sparsification to develop a kernelized SFA algorithm which provides a powerful function class for large data sets. Sparsity is achieved by a novel matching pursuit approach that can be applied to other tasks as well. For small data sets, however, the kernel SFA approach leads to over-fitting and numerical instabilities. To enforce a stable solution, we introduce regularization to the SFA objective. We hypothesize that our algorithm generates a feature space that resembles a Fourier basis in the unknown space of latent variables underlying a given real-world time series. We evaluate this hypothesis at the example of a vowel classification task in comparison to sparse kernel PCA. Our results show excellent classification accuracy and demonstrate the superiority of kernel SFA over kernel PCA in encoding latent variables.
引用
收藏
页码:67 / 86
页数:20
相关论文
共 34 条
[1]  
[Anonymous], 1961, PRINCIPLES NEURODYNA
[2]  
[Anonymous], 2000, P 17 INT C MACHINE L
[3]  
[Anonymous], 1973, Pattern Classification and Scene Analysis
[4]  
[Anonymous], 2006, Pattern recognition and machine learning
[5]  
Assmann Peter F., 2008, Canadian Acoustics, V36, P148
[6]   SELF-ORGANIZING NEURAL NETWORK THAT DISCOVERS SURFACES IN RANDOM-DOT STEREOGRAMS [J].
BECKER, S ;
HINTON, GE .
NATURE, 1992, 355 (6356) :161-163
[7]   Slow feature analysis yields a rich repertoire of complex cell properties [J].
Berkes, P ;
Wiskott, L .
JOURNAL OF VISION, 2005, 5 (06) :579-602
[8]  
Berkes P., 2005, 4104 COGN SCI EPRINT
[9]  
Böhmer W, 2011, LECT NOTES ARTIF INT, V6911, P235, DOI 10.1007/978-3-642-23780-5_25
[10]  
Bray A., 2002, NEURAL INFORM PROCES, P253