To Talk or to Work: Dynamic Batch Sizes Assisted Time Efficient Federated Learning Over Future Mobile Edge Devices

被引:20
作者
Shi, Dian [1 ]
Li, Liang [2 ]
Wu, Maoqiang [3 ]
Shu, Minglei [4 ]
Yu, Rong [2 ]
Pan, Miao [1 ]
Han, Zhu [5 ]
机构
[1] Univ Houston, Elect & Comp Engn Dept, Houston, TX USA
[2] Beijing Univ Posts & Telecommun, Sch Comp Sci, Natl Pilot Software Engn Sch, Beijing, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou, Peoples R China
[4] Qilu Univ Technol, Shandong Artificial Intelligence Inst, Shandong Acad Sci, Jinan, Peoples R China
[5] Univ Houston, Dept Elect & Comp Engn, Houston, TX USA
基金
美国国家科学基金会; 中国国家自然科学基金;
关键词
Federated learning; dynamic batch sizes; future mobile edge devices; on-GPU computing; OPTIMIZATION; 5G;
D O I
10.1109/TWC.2022.3189320
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The coupling of federated learning (FL) and multi-access edge computing (MEC) has the potential to foster numerous applications. However, it poses great challenges to train FL fast enough with limited communication and computing resources of mobile edge devices. Motivated by recent development in ultra fast wireless transmissions and promising advances in artificial intelligence (AI) computing hardware of mobile devices, in this paper, we propose a time efficient FL over future mobile edge devices, called dynamic batch sizes assisted federated learning (DBFL) with convergence guarantee. The DBFL allows batch sizes to increase dynamically during training, which can unleash the computing potential of GPU's parallelism for on- device training and effectively leverage the fast wireless transmissions (WiFi-6, 5G, 6G, etc.) of mobile edge devices. Furthermore, based on the derived DBFL's convergence bound, we develop a batch size control scheme to minimize the total time consumption of FL over mobile edge devices, which trade-offs the "talking ", i.e., communication time, and "working ", i.e., computing time, by adjusting the incremental factor appropriately. Extensive simulations are conducted to validate the effectiveness of our proposed DBFL algorithm and demonstrate that our scheme outperforms existing time efficient FL approaches in terms of the total time consumption in various settings.
引用
收藏
页码:11038 / 11050
页数:13
相关论文
共 35 条
[1]   Power and Performance Characterization and Modeling of GPU-Accelerated Systems [J].
Abe, Yuki ;
Inoue, Koji ;
Sasaki, Hiroshi ;
Edahiro, Masato ;
Kato, Shinpei ;
Peres, Martin .
2014 IEEE 28TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, 2014,
[2]  
Apple S. T., Hey Siri: An on-Device DNN-Powered Voice Trigger for Apple's Personal Assistant
[3]   Optimization Methods for Large-Scale Machine Learning [J].
Bottou, Leon ;
Curtis, Frank E. ;
Nocedal, Jorge .
SIAM REVIEW, 2018, 60 (02) :223-311
[4]   Federated learning of predictive models from federated Electronic Health Records [J].
Brisimi, Theodora S. ;
Chen, Ruidi ;
Mela, Theofanie ;
Olshevsky, Alex ;
Paschalidis, Ioannis Ch. ;
Shi, Wei .
INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2018, 112 :59-67
[5]   Convergence Time Optimization for Federated Learning Over Wireless Networks [J].
Chen, Mingzhe ;
Poor, H. Vincent ;
Saad, Walid ;
Cui, Shuguang .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (04) :2457-2471
[6]   Convergence Time Minimization of Federated Learning over Wireless Networks [J].
Chen, Mingzhe ;
Poor, H. Vincent ;
Saad, Walid ;
Cui, Shuguang .
ICC 2020 - 2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2020,
[7]  
De S, 2017, PR MACH LEARN RES, V54, P1504
[8]  
Devarakonda A., 2017, arXiv
[9]  
Goyal P, 2018, Arxiv, DOI arXiv:1706.02677
[10]  
Hard A, 2019, Arxiv, DOI arXiv:1811.03604