A Comprehensively Tight Analysis of Gradient Descent for PCA

被引:0
|
作者
Xu, Zhiqiang [1 ]
Li, Ping
机构
[1] Baidu Res, Cognit Comp Lab, 10 Xibeiwang East Rd, Beijing 100193, Peoples R China
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021) | 2021年 / 34卷
关键词
OPTIMIZATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the Riemannian gradient method for PCA on which a crucial fact is that despite the simplicity of the considered setting, i.e., deterministic version of Krasulina's method, the convergence rate has not been well-understood yet. In this work, we provide a general tight analysis for the gap-dependent rate at O(1/Delta log 1/is an element of) that holds for any real symmetric matrix. More importantly, when the gap is significantly smaller than the target accuracy. on the objective suboptimality of the final solution, the rate of this type is actually not tight any more, which calls for a worst-case rate. We further give the first worst-case analysis that achieves a rate of convergence at O(1/is an element of log 1/is an element of. The two analyses naturally roll out a comprehensively tight convergence rate at O(1/max{Delta,}log 1/is an element of). Particularly, our gap-dependent analysis suggests a new promising learning rate for stochastic variance reduced PCA algorithms. Experiments are conducted to confirm our findings as well.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] On Convergence of Gradient Descent Ascent: A Tight Local Analysis
    Li, Haochuan
    Farnia, Farzan
    Das, Subhro
    Jadbabaie, Ali
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] Convergence of Stochastic Gradient Descent for PCA
    Shamir, Ohad
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [3] A Tight Convergence Analysis for Stochastic Gradient Descent with Delayed Updates
    Arjevani, Yossi
    Shamir, Ohad
    Srebro, Nathan
    ALGORITHMIC LEARNING THEORY, VOL 117, 2020, 117 : 111 - 132
  • [4] Fast Algorithms for Robust PCA via Gradient Descent
    Yi, Xinyang
    Park, Dohyung
    Chen, Yudong
    Caramanis, Constantine
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [5] Tight Risk Bounds for Gradient Descent on Separable Data
    Schliserman, Matan
    Koren, Tomer
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Tight Convergence Rate of Gradient Descent for Eigenvalue Computation
    Ding, Qinghua
    Zhou, Kaiwen
    Cheng, James
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3276 - 3282
  • [7] Tight analyses for non-smooth stochastic gradient descent
    Harvey, Nicholas J. A.
    Liaw, Christopher
    Plan, Yaniv
    Randhawa, Sikander
    CONFERENCE ON LEARNING THEORY, VOL 99, 2019, 99
  • [8] Image Alignment by Online Robust PCA via Stochastic Gradient Descent
    Song, Wenjie
    Zhu, Jianke
    Li, Yang
    Chen, Chun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2016, 26 (07) : 1241 - 1250
  • [9] Analysis of the Whiplash gradient descent dynamics
    Bhattacharjee, Subhransu S.
    Petersen, Ian R.
    ASIAN JOURNAL OF CONTROL, 2025, 27 (01) : 27 - 40
  • [10] Learning to learn by gradient descent by gradient descent
    Andrychowicz, Marcin
    Denil, Misha
    Colmenarejo, Sergio Gomez
    Hoffman, Matthew W.
    Pfau, David
    Schaul, Tom
    Shillingford, Brendan
    de Freitas, Nando
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29