The Trimmed Lasso: Sparse Recovery Guarantees and Practical Optimization by the Generalized Soft-Min Penalty

被引:11
作者
Amir, Tal [1 ]
Basri, Ronen [1 ]
Nadler, Boaz [1 ]
机构
[1] Weizmann Inst Sci, Dept Comp Sci & Appl Math, IL-7610001 Rehovot, Israel
来源
SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE | 2021年 / 3卷 / 03期
关键词
best subset selection; sparse recovery; trimmed lasso; nonseparable penalty; NONCONCAVE PENALIZED LIKELIHOOD; ROBUST UNCERTAINTY PRINCIPLES; RESTRICTED ISOMETRY PROPERTY; REWEIGHTED LEAST-SQUARES; VARIABLE SELECTION; SIGNAL RECOVERY; LINEAR-MODELS; ALGORITHMS; MINIMIZATION; RECONSTRUCTION;
D O I
10.1137/20M1330634
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We present a new approach to solve the sparse approximation or best subset selection problem, namely find a k-sparse vector x is an element of R-d that minimizes the l(2) residual parallel to Ax - y parallel to(2). We consider a regularized approach, whereby this residual is penalized by the nonconvex trimmed lasso, defined as the l(1)-norm of x excluding its k largest-magnitude entries. We prove that the trimmed lasso has several appealing theoretical properties, and in particular derive sparse recovery guarantees assuming successful optimization of the penalized objective. Next, we show empirically that directly optimizing this objective can be quite challenging. Instead, we propose a surrogate for the trimmed lasso, called the generalized soft-min. This penalty smoothly interpolates between the classical lasso and the trimmed lasso, while taking into account all possible k-sparse patterns. The generalized soft-min penalty involves summation over ((d)(k)) terms, yet we derive a polynomial-time algorithm to compute it. This, in turn, yields a practical method for the original sparse approximation problem. Via simulations, we demonstrate its competitive performance compared to current state of the art.
引用
收藏
页码:900 / 929
页数:30
相关论文
共 94 条
[51]   REGRESSIONS BY LEAPS AND BOUNDS [J].
FURNIVAL, GM ;
WILSON, RW .
TECHNOMETRICS, 1974, 16 (04) :499-511
[52]   DC formulations and algorithms for sparse optimization problems [J].
Gotoh, Jun-ya ;
Takeda, Akiko ;
Tono, Katsuya .
MATHEMATICAL PROGRAMMING, 2018, 169 (01) :141-176
[53]  
Hamerly G., 2002, Proceedings of the Eleventh International Conference on Information and Knowledge Management. CIKM 2002, P600, DOI 10.1145/584792.584890
[54]  
Hastie T., 2015, STAT LEARNING SPARSI, DOI [10.1201/b18401, DOI 10.1201/B18401]
[55]   Fast Best Subset Selection: Coordinate Descent and Local Combinatorial Optimization Algorithms [J].
Hazimeh, Hussein ;
Mazumder, Rahul .
OPERATIONS RESEARCH, 2020, 68 (05) :1517-1537
[56]   Two-level l1 minimization for compressed sensing [J].
Huang, Xiaolin ;
Liu, Yipeng ;
Shi, Lei ;
Van Huffel, Sabine ;
Suykens, Johan A. K. .
SIGNAL PROCESSING, 2015, 108 :459-475
[57]   Variable selection using MM algorithms [J].
Hunter, DR ;
Li, RZ .
ANNALS OF STATISTICS, 2005, 33 (04) :1617-1642
[58]   A tutorial on MM algorithms [J].
Hunter, DR ;
Lange, K .
AMERICAN STATISTICIAN, 2004, 58 (01) :30-37
[59]  
Kincaid D., 1996, Numerical Analysis: Mathematics of Science Computing
[60]   IMPROVED ITERATIVELY REWEIGHTED LEAST SQUARES FOR UNCONSTRAINED SMOOTHED lq MINIMIZATION [J].
Lai, Ming-Jun ;
Xu, Yangyang ;
Yin, Wotao .
SIAM JOURNAL ON NUMERICAL ANALYSIS, 2013, 51 (02) :927-957