Linear Regression With Distributed Learning: A Generalization Error Perspective

被引:5
作者
Hellkvist, Martin [1 ]
Ozcelikkale, Ayca [1 ]
Ahlen, Anders [1 ]
机构
[1] Uppsala Univ, Dept Elect Engn, S-75237 Uppsala, Sweden
基金
瑞典研究理事会;
关键词
Distance learning; Computer aided instruction; Training; Data models; Training data; Distributed databases; Numerical models; Distributed estimation; distributed optimization; supervised learning; generalization error; networked systems; OPTIMIZATION; ALGORITHMS;
D O I
10.1109/TSP.2021.3106441
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Distributed learning provides an attractive framework for scaling the learning task by sharing the computational load over multiple nodes in a network. Here, we investigate the performance of distributed learning for large-scale linear regression where the model parameters, i.e., the unknowns, are distributed over the network. We adopt a statistical learning approach. In contrast to works that focus on the performance on the training data, we focus on the generalization error, i.e., the performance on unseen data. We provide high-probability bounds on the generalization error for both isotropic and correlated Gaussian data as well as sub-gaussian data. These results reveal the dependence of the generalization performance on the partitioning of the model over the network. In particular, our results show that the generalization error of the distributed solution can be substantially higher than that of the centralized solution even when the error on the training data is at the same level for both the centralized and distributed approaches. Our numerical results illustrate the performance with both real-world image data as well as synthetic data.
引用
收藏
页码:5479 / 5495
页数:17
相关论文
共 50 条
  • [1] Distributed Continual Learning With CoCoA in High-Dimensional Linear Regression
    Hellkvist, Martin
    Ozcelikkale, Ayca
    Ahlen, Anders
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1015 - 1031
  • [2] Generalization Error for Linear Regression under Distributed Learning
    Hellkvist, Martin
    Ozcelikkale, Ayca
    Ahlen, Anders
    PROCEEDINGS OF THE 21ST IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC2020), 2020,
  • [3] Generalization Error Matters in Decentralized Learning Under Byzantine Attacks
    Ye, Haoxiang
    Ling, Qing
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 843 - 857
  • [4] Lower Bounds on the Generalization Error of Nonlinear Learning Models
    Seroussi, Inbar
    Zeitouni, Ofer
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (12) : 7956 - 7970
  • [5] Generalization error estimation for non-linear learning methods
    Sugiyama, Masashi
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2007, E90A (07) : 1496 - 1499
  • [6] Rate-Distortion Theoretic Bounds on Generalization Error for Distributed Learning
    Sefidgaran, Milad
    Chor, Romain
    Zaidi, Abdellatif
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [7] Gradient Inversion of Text-Modal Data in Distributed Learning
    Ye, Zipeng
    Luo, Wenjian
    Zhou, Qi
    Tang, Yubo
    Zhu, Zhenqian
    Shi, Yuhui
    Jia, Yan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2025, 20 : 928 - 943
  • [8] Sparsity-Aware Distributed Learning for Gaussian Processes With Linear Multiple Kernel
    Suwandi, Richard Cornelius
    Lin, Zhidi
    Yin, Feng
    Wang, Zhiguo
    Theodoridis, Sergios
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025,
  • [9] Distributed Sparse Linear Regression
    Mateos, Gonzalo
    Bazerque, Juan Andres
    Giannakis, Georgios B.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2010, 58 (10) : 5262 - 5276
  • [10] Communication-Efficient Distributed Learning: An Overview
    Cao, Xuanyu
    Basar, Tamer
    Diggavi, Suhas
    Eldar, Yonina C.
    Letaief, Khaled B.
    Poor, H. Vincent
    Zhang, Junshan
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2023, 41 (04) : 851 - 873