Deterministic Gradient-Descent Learning of Linear Regressions: Adaptive Algorithms, Convergence Analysis and Noise Compensation

被引:0
|
作者
Liu, Kang-Zhi [1 ]
Gan, Chao [2 ]
机构
[1] Chiba Univ, Dept Elect & Elect Engn, Chiba 2638522, Japan
[2] China Univ Geosci, Sch Automat, Wuhan 430074, Peoples R China
关键词
Linear regression; gradient descent; adaptive learning rate; weight convergence; noise compensation; LMS;
D O I
10.1109/TPAMI.2024.3399312
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Weight learning forms a basis for the machine learning and numerous algorithms have been adopted up to date. Most of the algorithms were either developed in the stochastic framework or aimed at minimization of loss or regret functions. Asymptotic convergence of weight learning, vital for good output prediction, was seldom guaranteed for online applications. Since linear regression is the most fundamental component in machine learning, we focus on this model in this paper. Aiming at online applications, a deterministic analysis method is developed based on LaSalle's invariance principle. Convergence conditions are derived for both the first-order and the second-order learning algorithms, without resorting to any stochastic argument. Moreover, the deterministic approach makes it easy to analyze the noise influence. Specifically, adaptive hyperparameters are derived in this framework and their tuning rules disclosed for the compensation of measurement noise. Comparison with four most popular algorithms validates that this approach has a higher learning capability and is quite promising in enhancing the weight learning performance.
引用
收藏
页码:7867 / 7877
页数:11
相关论文
共 38 条
  • [1] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    arXiv, 2019,
  • [2] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [3] Gradient-descent based window optimization for linear prediction analysis
    Chu, WC
    2003 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL I, PROCEEDINGS: SPEECH PROCESSING I, 2003, : 460 - 463
  • [4] Convergence analysis of gradient descent stochastic algorithms
    Shapiro, A
    Wardi, Y
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 1996, 91 (02) : 439 - 454
  • [5] A gradient-descent iterative learning control algorithm for a non-linear system
    He, Zhiying
    Pu, Hongji
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2025, 47 (02) : 342 - 351
  • [6] Convergence of gradient descent for learning linear neural networks
    Nguegnang, Gabin Maxime
    Rauhut, Holger
    Terstiege, Ulrich
    ADVANCES IN CONTINUOUS AND DISCRETE MODELS, 2024, 2024 (01):
  • [7] Analysis of the Gradient-Descent Total Least-Squares Adaptive Filtering Algorithm
    Arablouei, Reza
    Werner, Stefan
    Dogancay, Kutluyil
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2014, 62 (05) : 1256 - 1264
  • [8] Impact of Mathematical Norms on Convergence of Gradient Descent Algorithms for Deep Neural Networks Learning
    Cai, Linzhe
    Yu, Xinghuo
    Li, Chaojie
    Eberhard, Andrew
    Lien Thuy Nguyen
    Chuong Thai Doan
    AI 2022: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13728 : 131 - 144
  • [9] Convergence Analysis of Distributed Gradient Descent Algorithms With One and Two Momentum Terms
    Liu, Bing
    Chai, Li
    Yi, Jingwen
    IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (03) : 1511 - 1522
  • [10] ANALYSIS OF GRADIENT DESCENT LEARNING ALGORITHMS FOR MULTILAYER FEEDFORWARD NEURAL NETWORKS
    GUO, H
    GELFAND, SB
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS, 1991, 38 (08): : 883 - 894