GRASSMANNIAN DIFFUSION MAPS-BASED DIMENSION REDUCTION AND CLASSIFICATION FOR HIGH-DIMENSIONAL DATA

被引:8
作者
Dos Santos, Ketson R. [1 ]
Giovanis, Dimitrios G. [1 ]
Shields, Michael D. [1 ]
机构
[1] Johns Hopkins Univ, Dept Civil & Syst Engn, Baltimore, MD 21218 USA
关键词
Grassmann manifold; diffusion maps; dimension reduction; data classification; face recognition; SUBSPACES; EIGENMAPS;
D O I
10.1137/20M137001X
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This work introduces the Grassmannian diffusion maps (GDMaps), a novel nonlinear dimensionality reduction technique that defines the affinity between points through their representation as low-dimensional subspaces corresponding to points on the Grassmann manifold. The method is designed for applications, such as image recognition and data-based classification of constrained high-dimensional data where each data point itself is a high-dimensional object (i.e., a large matrix) that can be compactly represented in a lower-dimensional subspace. The GDMaps is composed of two stages. The first is a pointwise linear dimensionality reduction wherein each high-dimensional object is mapped onto the Grassmann manifold representing the low-dimensional subspace on which it resides. The second stage is a multipoint nonlinear kernel-based dimension reduction using diffusion maps to identify the subspace structure of the points on the Grassmann manifold. To this end, an appropriate Grassmannian kernel is used to construct the transition matrix of a random walk on a graph connecting points on the Grassmann manifold. Spectral analysis of the transition matrix yields low-dimensional Grassmannian diffusion coordinates embedding the data into a low-dimensional reproducing kernel Hilbert space. Further, a novel data classification/recognition technique is developed based on the construction of an overcomplete dictionary of reduced dimension whose atoms are given by the Grassmannian diffusion coordinates. Three examples are considered. First, a "toy" example shows that the GDMaps can identify an appropriate parametrization of structured points on the unit sphere. The second example demonstrates the ability of the GDMaps to revealing the intrinsic subspace structure of high-dimensional random field data. In the last example, a face recognition problem is solved considering face images subject to varying illumination conditions, changes in face expressions, and occurrence of occlusions. The technique presented high recognition rates (i.e., 95% in the best case) using a fraction of the data required by conventional methods.
引用
收藏
页码:B250 / B274
页数:25
相关论文
共 64 条
  • [1] On the largest principal angle between random subspaces
    Absil, PA
    Edelman, A
    Koev, P
    [J]. LINEAR ALGEBRA AND ITS APPLICATIONS, 2006, 414 (01) : 288 - 294
  • [2] Aggarwal CC, 2001, LECT NOTES COMPUT SC, V1973, P420
  • [3] [Anonymous], 2019, P INT C LEARN REPR
  • [4] [Anonymous], 2006, 2006 IEEE COMPUTER S, DOI DOI 10.1109/CVPR.2006.50
  • [5] [Anonymous], 2008, P ADV NEUR INF PROC
  • [6] Auslander L., 2012, INTRO DIFFERENTIABLE
  • [7] Balasubramanian M, 2002, SCIENCE, V295
  • [8] Low-Dimensional Models for Dimensionality Reduction and Signal Recovery: A Geometric Perspective
    Baraniuk, Richard G.
    Cevher, Volkan
    Wakin, Michael B.
    [J]. PROCEEDINGS OF THE IEEE, 2010, 98 (06) : 959 - 971
  • [9] Laplacian eigenmaps for dimensionality reduction and data representation
    Belkin, M
    Niyogi, P
    [J]. NEURAL COMPUTATION, 2003, 15 (06) : 1373 - 1396
  • [10] Bendokat T., 2020, GRASSMANN MANIFOLD H