Flat minima generalize for low-rank matrix recovery

被引:0
|
作者
Ding, Lijun [1 ]
Drusvyatskiy, Dmitriy [2 ]
Fazel, Maryam [3 ]
Harchaoui, Zaid [4 ]
机构
[1] Univ Wisconsin, Wisconsin Inst Discovery, Madison, WI 53795 USA
[2] Univ Washington, Dept Math, Seattle, WA 98195 USA
[3] Univ Washington, Elect & Comp Engn Dept, Seattle, WA 98195 USA
[4] Univ Washington, Dept Stat, Seattle, WA 98195 USA
关键词
Flat minimizers; overparameterization; low rank matrix recovery; nuclear norm; matrix sensing; bilinear sensing; matrix completion; covariance matrix estimation; robust PCA; neural network; NEURAL-NETWORKS; INCOHERENCE;
D O I
10.1093/imaiai/iaae009
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Empirical evidence suggests that for a variety of overparameterized nonlinear models, most notably in neural network training, the growth of the loss around a minimizer strongly impacts its performance. Flat minima-those around which the loss grows slowly-appear to generalize well. This work takes a step towards understanding this phenomenon by focusing on the simplest class of overparameterized nonlinear models: those arising in low-rank matrix recovery. We analyse overparameterized matrix and bilinear sensing, robust principal component analysis, covariance matrix estimation and single hidden layer neural networks with quadratic activation functions. In all cases, we show that flat minima, measured by the trace of the Hessian, exactly recover the ground truth under standard statistical assumptions. For matrix completion, we establish weak recovery, although empirical evidence suggests exact recovery holds here as well. We complete the paper with synthetic experiments that illustrate our findings.
引用
收藏
页数:40
相关论文
共 50 条
  • [1] On the Absence of Spurious Local Minima in Nonlinear Low-Rank Matrix Recovery Problems
    Bi, Yingjie
    Lavaei, Javad
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 379 - +
  • [2] Quantization for low-rank matrix recovery
    Lybrand, Eric
    Saab, Rayan
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2019, 8 (01) : 161 - 180
  • [3] Sensitivity of low-rank matrix recovery
    Breiding, Paul
    Vannieuwenhoven, Nick
    NUMERISCHE MATHEMATIK, 2022, 152 (04) : 725 - 759
  • [4] Sensitivity of low-rank matrix recovery
    Paul Breiding
    Nick Vannieuwenhoven
    Numerische Mathematik, 2022, 152 : 725 - 759
  • [5] Maximum Entropy Low-Rank Matrix Recovery
    Mak, Simon
    Xie, Yao
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2018, 12 (05) : 886 - 901
  • [6] NONCONVEX ROBUST LOW-RANK MATRIX RECOVERY
    Li, Xiao
    Zhu, Zhihui
    So, Anthony Man-Cho
    Vidal, Rene
    SIAM JOURNAL ON OPTIMIZATION, 2020, 30 (01) : 660 - 686
  • [7] Low-Rank Matrix Recovery with Unknown Correspondence
    Tang, Zhiwei
    Chang, Tsung-Hui
    Ye, Xiaojing
    Zha, Hongyuan
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 2111 - 2122
  • [8] LOW-RANK MATRIX RECOVERY OF DYNAMIC EVENTS
    Asif, M. Salman
    2017 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2017), 2017, : 1215 - 1219
  • [9] Matrix recovery with implicitly low-rank data
    Xie, Xingyu
    Wu, Jianlong
    Liu, Guangcan
    Wang, Jun
    NEUROCOMPUTING, 2019, 334 : 219 - 226
  • [10] LOW-RANK MATRIX RECOVERY IN POISSON NOISE
    Cao, Yang
    Xie, Yao
    2014 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2014, : 384 - 388