The Cost of Training Machine Learning Models Over Distributed Data Sources

被引:10
作者
Guerra, Elia [1 ]
Wilhelmi, Francesc [2 ]
Miozzo, Marco [1 ]
Dini, Paolo [1 ]
机构
[1] Ctr Tecnol Telecomunicac Catalunya CTTC CERCA, Sustainable Artificial Intelligence SAI, Castelldefels 08860, Spain
[2] Nokia Bell Labs, Radio Syst Res, D-70469 Stuttgart, Germany
来源
IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY | 2023年 / 4卷
基金
欧盟地平线“2020”;
关键词
Servers; Data models; Training; Federated learning; Security; Energy consumption; Blockchains; Blockchain; decentralized learning; edge computing; energy efficiency; federated learning; machine learning; EDGE; BLOCKCHAIN;
D O I
10.1109/OJCOMS.2023.3274394
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Federated learning is one of the most appealing alternatives to the standard centralized learning paradigm, allowing a heterogeneous set of devices to train a machine learning model without sharing their raw data. However, it requires a central server to coordinate the learning process, thus introducing potential scalability and security issues. In the literature, server-less federated learning approaches like gossip federated learning and blockchain-enabled federated learning have been proposed to mitigate these issues. In this work, we propose a complete overview of these three techniques, proposing a comparison according to an integral set of performance indicators, including model accuracy, time complexity, communication overhead, convergence time, and energy consumption. An extensive simulation campaign permits to draw a quantitative analysis considering both feedforward and convolutional neural network models. Results show that gossip federated learning and standard federated solution are able to reach a similar level of accuracy, and their energy consumption is influenced by the machine learning model adopted, the software library, and the hardware used. Differently, blockchain-enabled federated learning represents a viable solution for implementing decentralized learning with a higher level of security, at the cost of an extra energy usage and data sharing. Finally, we identify open issues on the two decentralized federated learning implementations and provide insights on potential extensions and possible research directions on this new research field.
引用
收藏
页码:1111 / 1126
页数:16
相关论文
共 65 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   Estimating Energy Consumption of Cloud, Fog, and Edge Computing Infrastructures [J].
Ahvar, Ehsan ;
Orgerie, Anne-Cecile ;
Lebre, Adrien .
IEEE TRANSACTIONS ON SUSTAINABLE COMPUTING, 2022, 7 (02) :277-288
[3]  
Aledhari M, 2020, IEEE ACCESS, V8, P140699, DOI [10.1109/ACCESS.2020.3013541, 10.1109/access.2020.3013541]
[4]   BlockSim: An Extensible Simulation Tool for Blockchain Systems [J].
Alharby, Maher ;
van Moorsel, Aad .
FRONTIERS IN BLOCKCHAIN, 2020, 3
[5]  
[Anonymous], 2021, Ericsson mobility report
[6]  
[Anonymous], EMNIST TENSFLOW FED
[7]  
[Anonymous], KERAS PYTHON DEEP LE
[8]  
[Anonymous], 2018, Ai and compute
[9]   FLChain: A Blockchain for Auditable Federated Learning with Trust and Incentive [J].
Bao, Xianglin ;
Su, Cheng ;
Xiong, Yan ;
Huang, Wenchao ;
Hu, Yifei .
5TH INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING AND COMMUNICATIONS (BIGCOM 2019), 2019, :151-159
[10]   Decentralized federated learning for extended sensing in 6G connected vehicles [J].
Barbieri, Luca ;
Savazzi, Stefano ;
Brambilla, Mattia ;
Nicoli, Monica .
VEHICULAR COMMUNICATIONS, 2022, 33