A global two-stage algorithm for non-convex penalized high-dimensional linear regression problems

被引:1
|
作者
Li, Peili [1 ]
Liu, Min [2 ]
Yu, Zhou [1 ]
机构
[1] East China Normal Univ, KLATASDS MOE, Sch Stat, Shanghai 200062, Peoples R China
[2] Wuhan Univ, Sch Math & Stat, Wuhan 430072, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
High-dimensional linear regression; Global convergence; Two-stage algorithm; Primal dual active set with continuation algorithm; Difference of convex functions; COORDINATE DESCENT ALGORITHMS; ACTIVE SET ALGORITHM; VARIABLE SELECTION; LIKELIHOOD;
D O I
10.1007/s00180-022-01249-w
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
By the asymptotic oracle property, non-convex penalties represented by minimax concave penalty (MCP) and smoothly clipped absolute deviation (SCAD) have attracted much attentions in high-dimensional data analysis, and have been widely used in signal processing, image restoration, matrix estimation, etc. However, in view of their non-convex and non-smooth characteristics, they are computationally challenging. Almost all existing algorithms converge locally, and the proper selection of initial values is crucial. Therefore, in actual operation, they often combine a warm-starting technique to meet the rigid requirement that the initial value must be sufficiently close to the optimal solution of the corresponding problem. In this paper, based on the DC (difference of convex functions) property of MCP and SCAD penalties, we aim to design a global two-stage algorithm for the high-dimensional least squares linear regression problems. A key idea for making the proposed algorithm to be efficient is to use the primal dual active set with continuation (PDASC) method to solve the corresponding sub-problems. Theoretically, we not only prove the global convergence of the proposed algorithm, but also verify that the generated iterative sequence converges to a d-stationary point. In terms of computational performance, the abundant research of simulation and real data show that the algorithm in this paper is superior to the latest SSN method and the classic coordinate descent (CD) algorithm for solving non-convex penalized high-dimensional linear regression problems.
引用
收藏
页码:871 / 898
页数:28
相关论文
共 50 条
  • [31] Lasso penalized model selection criteria for high-dimensional multivariate linear regression analysis
    Katayama, Shota
    Imori, Shinpei
    JOURNAL OF MULTIVARIATE ANALYSIS, 2014, 132 : 138 - 150
  • [32] A two-stage sequential conditional selection approach to sparse high-dimensional multivariate regression models
    Chen, Zehua
    Jiang, Yiwei
    ANNALS OF THE INSTITUTE OF STATISTICAL MATHEMATICS, 2020, 72 (01) : 65 - 90
  • [33] Two limit theorems for the high-dimensional two-stage contact process
    Xue, Xiaofeng
    ALEA-LATIN AMERICAN JOURNAL OF PROBABILITY AND MATHEMATICAL STATISTICS, 2020, 17 (02): : 825 - 855
  • [34] Two-stage extreme learning machine for high-dimensional data
    Liu, Peng
    Huang, Yihua
    Meng, Lei
    Gong, Siyuan
    Zhang, Guopeng
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2016, 7 (05) : 765 - 772
  • [35] Two-stage extreme learning machine for high-dimensional data
    Peng Liu
    Yihua Huang
    Lei Meng
    Siyuan Gong
    Guopeng Zhang
    International Journal of Machine Learning and Cybernetics, 2016, 7 : 765 - 772
  • [36] A multipurpose parallel genetic hybrid algorithm for non-linear non-convex programming problems
    Östermark, R
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2004, 152 (01) : 195 - 214
  • [37] Two-stage improved Grey Wolf optimization algorithm for feature selection on high-dimensional classification
    Chaonan Shen
    Kai Zhang
    Complex & Intelligent Systems, 2022, 8 : 2769 - 2789
  • [38] Two-stage improved Grey Wolf optimization algorithm for feature selection on high-dimensional classification
    Shen, Chaonan
    Zhang, Kai
    COMPLEX & INTELLIGENT SYSTEMS, 2022, 8 (04) : 2769 - 2789
  • [39] Online stochastic gradient descent on non-convex losses from high-dimensional inference
    Ben Arous, Gerard
    Gheissari, Reza
    Jagannath, Aukosh
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [40] Penalized least-squares estimation for regression coefficients in high-dimensional partially linear models
    Ni, Huey-Fan
    JOURNAL OF STATISTICAL PLANNING AND INFERENCE, 2012, 142 (02) : 379 - 389