Accelerated Path-Following Iterative Shrinkage Thresholding Algorithm With Application to Semiparametric Graph Estimation

被引:1
作者
Zhao, Tuo [1 ]
Liu, Han [2 ]
机构
[1] Johns Hopkins Univ, Dept Comp Sci, Baltimore, MD 21218 USA
[2] Princeton Univ, Dept Operat Res & Financial Engn, Princeton, NJ 08544 USA
基金
美国国家科学基金会;
关键词
Coordinate descent; High dimensions; Nonconvex regularization; Path-following optimization; Semiparametric graph estimation; Variable selection; PRECISION MATRIX ESTIMATION; NONCONVEX PENALIZED REGRESSION; COORDINATE DESCENT ALGORITHMS; GENERALIZED LINEAR-MODELS; VARIABLE SELECTION; COVARIANCE ESTIMATION; LIKELIHOOD-ESTIMATION; ADAPTIVE LASSO; REGULARIZATION; SPARSITY;
D O I
10.1080/10618600.2016.1164533
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We propose an accelerated path-following iterative shrinkage thresholding algorithm (APISTA) for solving high-dimensional sparse nonconvex learning problems. The main difference between APISTA and the path-following iterative shrinkage thresholding algorithm (PISTA) is that APISTA exploits an additional coordinate descent subroutine to boost the computational performance. Such a modification, though simple, has profound impact: APISTA not only enjoys the same theoretical guarantee as that of PISTA, that is, APISTA attains a linear rate of convergence to a unique sparse local optimum with good statistical properties, but also significantly outperforms PISTA in empirical benchmarks. As an application, we apply APISTA to solve a family of nonconvex optimization problems motivated by estimating sparse semiparametric graphical models. APISTA allows us to' obtain new statistical recovery results that do not exist in the existing literature. Thorough numerical results are provided to back up our theory.
引用
收藏
页码:1272 / 1296
页数:25
相关论文
共 67 条
  • [1] [Anonymous], 2009, P 26 ANN INT C MACH
  • [2] [Anonymous], 2005, Mathematical Surveys and Monographs
  • [3] [Anonymous], 2006, Journal of the Royal Statistical Society, Series B
  • [4] [Anonymous], 2014, Advances in Neural Information Processing Systems
  • [5] Banerjee O, 2008, J MACH LEARN RES, V9, P485
  • [6] Fast Gradient-Based Algorithms for Constrained Total Variation Image Denoising and Deblurring Problems
    Beck, Amir
    Teboulle, Marc
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2009, 18 (11) : 2419 - 2434
  • [7] COORDINATE DESCENT ALGORITHMS FOR NONCONVEX PENALIZED REGRESSION, WITH APPLICATIONS TO BIOLOGICAL FEATURE SELECTION
    Breheny, Patrick
    Huang, Jian
    [J]. ANNALS OF APPLIED STATISTICS, 2011, 5 (01) : 232 - 253
  • [8] A Constrained l1 Minimization Approach to Sparse Precision Matrix Estimation
    Cai, Tony
    Liu, Weidong
    Luo, Xi
    [J]. JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2011, 106 (494) : 594 - 607
  • [9] DENNIS J., 1983, Numerical Methods for Unconstrained Optimization and Nonlinear Equations, V16
  • [10] Fan JQ, 2012, J ROY STAT SOC B, V74, P745, DOI 10.1111/j.1467-9868.2012.01029.x