Geometry on Probability Spaces

被引:49
作者
Smale, Steve [2 ]
Zhou, Ding-Xuan [1 ]
机构
[1] City Univ Hong Kong, Dept Math, Kowloon, Hong Kong, Peoples R China
[2] Toyota Technol Inst, Chicago, IL 60637 USA
基金
美国国家科学基金会;
关键词
Learning theory; Reproducing kernel Hilbert space; Graph Laplacian; Dimensionality reduction; Integral operator; APPROXIMATION;
D O I
10.1007/s00365-009-9070-2
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Partial differential equations and the Laplacian operator on domains in Euclidean spaces have played a central role in understanding natural phenomena. However, this avenue has been limited in many areas where calculus is obstructed, as in singular spaces, and in function spaces of functions on a space X where X itself is a function space. Examples of the latter occur in vision and quantum field theory. In vision it would be useful to do analysis on the space of images and an image is a function on a patch. Moreover, in analysis and geometry, the Lebesgue measure and its counterpart on manifolds are central. These measures are unavailable in the vision example and even in learning theory in general. There is one situation where, in the last several decades, the problem has been studied with some success. That is when the underlying space is finite (or even discrete). The introduction of the graph Laplacian has been a major development in algorithm research and is certainly useful for unsupervised learning theory. The approach taken here is to take advantage of both the classical research and the newer graph theoretic ideas to develop geometry on probability spaces. This starts with a space X equipped with a kernel (like a Mercer kernel) which gives a topology and geometry; X is to be equipped as well with a probability measure. The main focus is on a construction of a (normalized) Laplacian, an associated heat equation, diffusion distance, etc. In this setting, the point estimates of calculus are replaced by integral quantities. One thinks of secants rather than tangents. Our main result bounds the error of an empirical approximation to this Laplacian on X.
引用
收藏
页码:311 / 323
页数:13
相关论文
共 25 条
[1]  
[Anonymous], 1997, REGIONAL C SERIES MA
[2]  
[Anonymous], 2007, Advances in Neural Information Processing Systems
[3]  
[Anonymous], 0723 UCLA CAM
[4]   THEORY OF REPRODUCING KERNELS [J].
ARONSZAJN, N .
TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY, 1950, 68 (MAY) :337-404
[5]   Laplacian eigenmaps for dimensionality reduction and data representation [J].
Belkin, M ;
Niyogi, P .
NEURAL COMPUTATION, 2003, 15 (06) :1373-1396
[6]  
BELKIN M, RANDOM ESTIMATES OPE
[7]  
Bhatia Rajendra, 1997, Matrix Analysis: Graduate Texts in Mathematics, V169
[8]   Statistical properties of kernel principal component analysis [J].
Blanchard, Gilles ;
Bousquet, Olivier ;
Zwald, Laurent .
MACHINE LEARNING, 2007, 66 (2-3) :259-294
[9]  
Bougleux S, 2007, LECT NOTES COMPUT SC, V4485, P128
[10]   Diffusion wavelets [J].
Coifman, Ronald R. ;
Maggioni, Mauro .
APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2006, 21 (01) :53-94