A Minibatch Proximal Stochastic Recursive Gradient Algorithm Using a Trust-Region-Like Scheme and Barzilai-Borwein Stepsizes

被引:7
|
作者
Yu, Tengteng [1 ]
Liu, Xin-Wei [2 ]
Dai, Yu-Hong [3 ,4 ]
Sun, Jie [5 ,6 ]
机构
[1] Hebei Univ Technol, Sch Artificial Intelligence, Tianjin 300401, Peoples R China
[2] Hebei Univ Technol, Inst Math, Tianjin 300401, Peoples R China
[3] Chinese Acad Sci, Acad Math & Syst Sci, State Key Lab Sci & Engn Comp LSEC, Beijing 100190, Peoples R China
[4] Univ Chinese Acad Sci, Sch Math Sci, Beijing 100049, Peoples R China
[5] Hebei Univ Technol, Inst Math, Tianjin 300401, Peoples R China
[6] Natl Univ Singapore, Sch Business, Singapore 119245, Singapore
关键词
Convergence; Convex functions; Risk management; Gradient methods; Learning systems; Sun; Barzilai-Borwein (BB) method; empirical risk minimization (ERM); proximal method; stochastic gradient; trust-region; MACHINE;
D O I
10.1109/TNNLS.2020.3025383
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of minimizing the sum of an average of a large number of smooth convex component functions and a possibly nonsmooth convex function that admits a simple proximal mapping. This class of problems arises frequently in machine learning, known as regularized empirical risk minimization (ERM). In this article, we propose mSRGTR-BB, a minibatch proximal stochastic recursive gradient algorithm, which employs a trust-region-like scheme to select stepsizes that are automatically computed by the Barzilai-Borwein method. We prove that mSRGTR-BB converges linearly in expectation for strongly and nonstrongly convex objective functions. With proper parameters, mSRGTR-BB enjoys a faster convergence rate than the state-of-the-art minibatch proximal variant of the semistochastic gradient method (mS2GD). Numerical experiments on standard data sets show that the performance of mSRGTR-BB is comparable to and sometimes even better than mS2GD with best-tuned stepsizes and is superior to some modern proximal stochastic gradient methods.
引用
收藏
页码:4627 / 4638
页数:12
相关论文
共 5 条
  • [1] A new inexact stochastic recursive gradient descent algorithm with Barzilai-Borwein step size in machine learning
    Yang, Yi-ming
    Wang, Fu-sheng
    Li, Jin-xiang
    Qin, Yuan-yuan
    NONLINEAR DYNAMICS, 2023, 111 (04) : 3575 - 3586
  • [2] Stochastic Variance Reduced Gradient Methods Using a Trust-Region-Like Scheme
    Tengteng Yu
    Xin-Wei Liu
    Yu-Hong Dai
    Jie Sun
    Journal of Scientific Computing, 2021, 87
  • [3] Stochastic Variance Reduced Gradient Methods Using a Trust-Region-Like Scheme
    Yu, Tengteng
    Liu, Xin-Wei
    Dai, Yu-Hong
    Sun, Jie
    JOURNAL OF SCIENTIFIC COMPUTING, 2021, 87 (01)
  • [4] A New Nonmonotone Trust Region Algorithm Based on Simulated Annealing and Barzilai-Borwein Methods for Unconstrained Optimization
    Mirzaei, Seyed Hamzeh
    Ashrafi, Ali
    ASIA-PACIFIC JOURNAL OF OPERATIONAL RESEARCH, 2025,
  • [5] A new inexact stochastic recursive gradient descent algorithm with Barzilai–Borwein step size in machine learning
    Yi-ming Yang
    Fu-sheng Wang
    Jin-xiang Li
    Yuan-yuan Qin
    Nonlinear Dynamics, 2023, 111 : 3575 - 3586