Dictionary-Based Low-Rank Approximations and the Mixed Sparse Coding Problem
被引:0
作者:
Cohen, Jeremy E.
论文数: 0引用数: 0
h-index: 0
机构:
Univ Lyon, INSA Lyon, UJM St Etienne, CNRS,Inserm,CREATIS UMR 5220,UCBL, Villeurbanne, FranceUniv Lyon, INSA Lyon, UJM St Etienne, CNRS,Inserm,CREATIS UMR 5220,UCBL, Villeurbanne, France
Cohen, Jeremy E.
[1
]
机构:
[1] Univ Lyon, INSA Lyon, UJM St Etienne, CNRS,Inserm,CREATIS UMR 5220,UCBL, Villeurbanne, France
Constrained tensor and matrix factorization models allow to extract interpretable patterns from multiway data. Therefore crafting efficient algorithms for constrained low-rank approximations is nowadays an important research topic. This work deals with columns of factor matrices of a low-rank approximation being sparse in a known and possibly overcomplete basis, a model coined as Dictionary-based Low-Rank Approximation (DLRA). While earlier contributions focused on finding factor columns inside a dictionary of candidate columns, i.e., one-sparse approximations, this work is the first to tackle DLRA with sparsity larger than one. I propose to focus on the sparse-coding subproblem coined Mixed Sparse-Coding (MSC) that emerges when solving DLRA with an alternating optimization strategy. Several algorithms based on sparse-coding heuristics (greedy methods, convex relaxations) are provided to solve MSC. The performance of these heuristics is evaluated on simulated data. Then, I show how to adapt an efficient MSC solver based on the LASSO to compute Dictionary-based Matrix Factorization and Canonical Polyadic Decomposition in the context of hyperspectral image processing and chemometrics. These experiments suggest that DLRA extends the modeling capabilities of low-rank approximations, helps reducing estimation variance and enhances the identifiability and interpretability of estimated factors.