Linear Regression With Distributed Learning: A Generalization Error Perspective

被引:5
作者
Hellkvist, Martin [1 ]
Ozcelikkale, Ayca [1 ]
Ahlen, Anders [1 ]
机构
[1] Uppsala Univ, Dept Elect Engn, S-75237 Uppsala, Sweden
基金
瑞典研究理事会;
关键词
Distance learning; Computer aided instruction; Training; Data models; Training data; Distributed databases; Numerical models; Distributed estimation; distributed optimization; supervised learning; generalization error; networked systems; OPTIMIZATION; ALGORITHMS;
D O I
10.1109/TSP.2021.3106441
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Distributed learning provides an attractive framework for scaling the learning task by sharing the computational load over multiple nodes in a network. Here, we investigate the performance of distributed learning for large-scale linear regression where the model parameters, i.e., the unknowns, are distributed over the network. We adopt a statistical learning approach. In contrast to works that focus on the performance on the training data, we focus on the generalization error, i.e., the performance on unseen data. We provide high-probability bounds on the generalization error for both isotropic and correlated Gaussian data as well as sub-gaussian data. These results reveal the dependence of the generalization performance on the partitioning of the model over the network. In particular, our results show that the generalization error of the distributed solution can be substantially higher than that of the centralized solution even when the error on the training data is at the same level for both the centralized and distributed approaches. Our numerical results illustrate the performance with both real-world image data as well as synthetic data.
引用
收藏
页码:5479 / 5495
页数:17
相关论文
共 50 条
  • [21] Lightweight Distributed Gaussian Process Regression for Online Machine Learning
    Yuan, Zhenyuan
    Zhu, Minghui
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (06) : 3928 - 3943
  • [22] Distributed learning for sketched kernel regression
    Lian, Heng
    Liu, Jiamin
    Fan, Zengyan
    NEURAL NETWORKS, 2021, 143 : 368 - 376
  • [23] Task-Aware Service Placement for Distributed Learning in Wireless Edge Networks
    Cong, Rong
    Zhao, Zhiwei
    Wang, Mengfan
    Min, Geyong
    Liu, Jiangshu
    Mo, Jiwei
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2025, 36 (04) : 731 - 744
  • [24] Distributed Online Learning With Multiple Kernels
    Hong, Songnam
    Chae, Jeongmin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (03) : 1263 - 1277
  • [25] Distributed Training of Deep Learning Models: A Taxonomic Perspective
    Langer, Matthias
    He, Zhen
    Rahayu, Wenny
    Xue, Yanbo
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (12) : 2802 - 2818
  • [26] GAN Supervised Seismic Data Reconstruction: An Enhanced Learning for Improved Generalization
    Goyes-Penafiel, Paul
    Suarez-Rodriguez, Leon
    Correa, Claudia V.
    Arguello, Henry
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [27] Distributed Ordinal Regression Over Networks
    Liu, Huan
    Tu, Jiankai
    Li, Chunguang
    IEEE ACCESS, 2021, 9 : 62493 - 62504
  • [28] Wireless Distributed Learning: A New Hybrid Split and Federated Learning Approach
    Liu, Xiaolan
    Deng, Yansha
    Mahmoodi, Toktam
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (04) : 2650 - 2665
  • [29] Individually Conditional Individual Mutual Information Bound on Generalization Error
    Zhou, Ruida
    Tian, Chao
    Liu, Tie
    IEEE TRANSACTIONS ON INFORMATION THEORY, 2022, 68 (05) : 3304 - 3316
  • [30] A Piecewise Linear Regression and Classification Algorithm With Application to Learning and Model Predictive Control of Hybrid Systems
    Bemporad, Alberto
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (06) : 3194 - 3209