Sparse Regression as a Sparse Eigenvalue Problem

被引:0
作者
Moghaddam, Baback [1 ]
Gruber, Amit [2 ]
Weiss, Yair [2 ]
Avidan, Shai [3 ]
机构
[1] CALTECH, Jet Prop Lab, Pasadena, CA 91125 USA
[2] Hebrew Univ Jerusalem, Jerusalem, Israel
[3] Adobe Syst Inc, Newton, MA USA
来源
2008 INFORMATION THEORY AND APPLICATIONS WORKSHOP | 2008年
基金
美国国家航空航天局;
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We extend the l(0)-norm "subspectral" algorithms developed for sparse-LDA [5] and sparse-PCA [6] to more general quadratic costs such as MSE in linear (or kernel) regression. The resulting "Sparse Least Squares" (SLS) problem is also NP-hard, by way of its equivalence to a rank-1 sparse eigenvalue problem. Specifically, for minimizing general quadratic cost functions we use a highly-efficient method for direct eigenvalue computation based on partitioned matrix inverse techniques that leads to x 10(3) speed-ups over standard eigenvalue decomposition. This increased efficiency mitigates the O(n(4)) complexity that limited the previous algorithms' utility for high-dimensional problems. Moreover, the new computation prioritizes the role of the less-myopic backward elimination stage which becomes even more efficient than forward selection. Similarly, branch-and-bound search for Exact Sparse Least Squares (ESLS) also benefits from partitioned matrix techniques. Our Greedy Sparse Least Squares (GSLS) algorithm generalizes Natarajan's algorithm [9] also known as Order-Recursive Matching Pursuit (ORMP). Specifically, the forward pass of GSLS is exactly equivalent to ORMP but is more efficient, and by including the backward pass, which only doubles the computation, we can achieve a lower MSE than ORMP. In experimental comparisons with LARS [3], forward-GSLS is shown to be not only more efficient and accurate but more flexible in terms of choice of regularization.
引用
收藏
页码:255 / +
页数:2
相关论文
共 12 条
[1]   Atomic decomposition by basis pursuit [J].
Chen, SSB ;
Donoho, DL ;
Saunders, MA .
SIAM JOURNAL ON SCIENTIFIC COMPUTING, 1998, 20 (01) :33-61
[2]  
DASPREMONT A, 2004, NEURAL INFORM PROCES, V17
[3]   Least angle regression - Rejoinder [J].
Efron, B ;
Hastie, T ;
Johnstone, I ;
Tibshirani, R .
ANNALS OF STATISTICS, 2004, 32 (02) :494-499
[4]  
Golub GH., 2013, Matrix Computations, DOI 10.56021/9781421407944
[5]  
MOGHADDAM B, 2007, INT C COMP VIS ICCV
[6]  
MOGHADDAM B, 2006, NEURAL INFORM PROCES, V18
[7]  
MOGHADDAM B, 2006, INT C MACH LEARN ICM
[8]   Some greedy learning algorithms for sparse regression and classification with Mercer kernels [J].
Nair, PB ;
Choudhury, A ;
Keane, AJ .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :781-801
[9]   SPARSE APPROXIMATE SOLUTIONS TO LINEAR-SYSTEMS [J].
NATARAJAN, BK .
SIAM JOURNAL ON COMPUTING, 1995, 24 (02) :227-234
[10]  
Nemhauser G. L., 1988, INTEGER COMBINATORIA