Constrained Stochastic Gradient Descent for Large-scale Least Squares Problem

被引:0
|
作者
Mu, Yang [1 ]
Ding, Wei [1 ]
Zhou, Tianyi [2 ]
Tao, Dacheng [2 ]
机构
[1] Univ Massachusetts, 100 Morrissey Blvd, Boston, MA 02125 USA
[2] Univ Technol Sydney, Ultimo, NSW 2007, Australia
来源
19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13) | 2013年
关键词
Stochastic optimization; Large-scale least squares; online learning; APPROXIMATION; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The least squares problem is one of the most important regression problems in statistics, machine learning and data mining. In this paper, we present the Constrained Stochastic Gradient Descent (CSGD) algorithm to solve the large-scale least squares problem. CSGD improves the Stochastic Gradient Descent (SGD) by imposing a provable constraint that the linear regression line passes through the mean point of all the data points. It results in the best regret bound o(logT), and fastest convergence speed among all first order approaches. Empirical studies justify the effectiveness of CSGD by comparing it with SGD and other state-of-the-art approaches. An example is also given to show how to use CSGD to optimize SGD based least squares problems to achieve a better performance.
引用
收藏
页码:883 / 891
页数:9
相关论文
共 50 条
  • [11] ON BOX-CONSTRAINED TOTAL LEAST SQUARES PROBLEM
    Xu, Zhuoyi
    Xia, Yong
    Han, Deren
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2020, 10 (04): : 439 - 449
  • [12] Adaptive Powerball Stochastic Conjugate Gradient for Large-Scale Learning
    Yang, Zhuang
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (06) : 1598 - 1606
  • [13] Large-scale machine learning with fast and stable stochastic conjugate gradient
    Yang, Zhuang
    COMPUTERS & INDUSTRIAL ENGINEERING, 2022, 173
  • [14] Inertial accelerated stochastic mirror descent for large-scale generalized tensor CP decomposition
    Liu, Zehui
    Wang, Qingsong
    Cui, Chunfeng
    Xia, Yong
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2025, : 201 - 233
  • [15] On Local Linear Convergence of Projected Gradient Descent for Unit-Modulus Least Squares
    Vu, Trung
    Raich, Raviv
    Fu, Xiao
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 3883 - 3897
  • [16] Stochastic Coordinate Descent Frank-Wolfe Algorithm for Large-Scale Biological Network Alignment
    Wang, Yijie
    Qian, Xiaoning
    2014 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2014, : 1410 - 1413
  • [17] A real unconstrained equivalent problem of the quaternion equality constrained weighted least squares problem
    Zhang, Fengxia
    Li, Ying
    Zhao, Jianli
    NUMERICAL ALGORITHMS, 2023, 94 (01) : 73 - 91
  • [18] AVERAGING RANDOM PROJECTION: A FAST ONLINE SOLUTION FOR LARGE-SCALE CONSTRAINED STOCHASTIC OPTIMIZATION
    Liu, Jialin
    Gu, Yuantao
    Wang, Mengdi
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 3586 - 3590
  • [19] An online conjugate gradient algorithm for large-scale data analysis in machine learning
    Xue, Wei
    Wan, Pengcheng
    Li, Qiao
    Zhong, Ping
    Yu, Gaohang
    Tao, Tao
    AIMS MATHEMATICS, 2021, 6 (02): : 1515 - 1537
  • [20] Towards Practical Large-Scale Randomized Iterative Least Squares Solvers through Uncertainty Quantification\ast
    Pritchad, Nathaniel
    Patel, Vivak
    SIAM-ASA JOURNAL ON UNCERTAINTY QUANTIFICATION, 2023, 11 (03) : 996 - 1024