An Efficient Distributed Machine Learning Framework in Wireless D2D Networks: Convergence Analysis and System Implementation

被引:4
作者
Cheng, Kun [1 ]
Guo, Fengxian [1 ]
Peng, Mugen [1 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
Network topology; Convergence; Frequency synchronization; Topology; Device-to-device communication; Training; Scheduling; Distributed machine learning; wireless D2D networks; convergence analysis; system implementation;
D O I
10.1109/TVT.2023.3234550
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Facing the heavy traffic burden and data privacy, distributed machine learning (DML) has been envisioned as a promising computing paradigm to enable edge intelligence by extracting and establishing models collaboratively in wireless networks. Unlike centralized methods, DML involves frequent local model exchanging among distributed devices, which confronts low efficiency referring to the convergence rate and delay. To enable edge intelligence, how to train DML efficiently has recently attracted extensive interest in wireless networks. Rather than over a fixed or centralized topology, we study the convergence and system implementation of DML over a wireless device-to-device (D2D) network in this paper. First, we introduce the DML training process and system model in this network, where the total delay in reaching convergence is minimized. Second, we analyze the convergence rate to figure out the special effects of synchronization frequency and network topology. To improve the efficiency of DML training, we propose a system implementation approach to reduce the convergence rate and delay of one iteration, where the network topology and synchronization frequency are set, followed by an optimal resource allocation policy. At last, we perform experiments with an image classification task. Simulation results indicate that our proposed D2D framework can effectively reduce training delay and promote computation efficiency by 39$\%$ in a heterogeneous environment, which is concordant with the theoretical analysis.
引用
收藏
页码:6723 / 6738
页数:16
相关论文
共 37 条
[1]  
Abad MSH, 2020, INT CONF ACOUST SPEE, P8866, DOI [10.1109/icassp40776.2020.9054634, 10.1109/ICASSP40776.2020.9054634]
[2]   AI-Enabled Secure Microservices in Edge Computing: Opportunities and Challenges [J].
Al-Doghman, Firas ;
Moustafa, Nour ;
Khalil, Ibrahim ;
Sohrabi, Nasrin ;
Tari, Zahir ;
Zomaya, Albert Y. .
IEEE TRANSACTIONS ON SERVICES COMPUTING, 2023, 16 (02) :1485-1504
[3]  
[Anonymous], 2018, Distributed Machine Learning-Theories, Algorithms and Systems
[4]   Toward On-Device Federated Learning: A Direct Acyclic Graph-Based Blockchain Approach [J].
Cao, Mingrui ;
Zhang, Long ;
Cao, Bin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (04) :2028-2042
[5]   Federated Learning Over Wireless Networks: Convergence Analysis and Resource Allocation [J].
Dinh, Canh T. ;
Tran, Nguyen H. ;
Nguyen, Minh N. H. ;
Hong, Choong Seon ;
Bao, Wei ;
Zomaya, Albert Y. ;
Gramoli, Vincent .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2021, 29 (01) :398-409
[6]   High-Dimensional Stochastic Gradient Quantization for Communication-Efficient Edge Learning [J].
Du, Yuqing ;
Yang, Sheng ;
Huang, Kaibin .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 :2128-2142
[7]   Link scheduling in wireless sensor networks: Distributed edge-coloring revisited [J].
Gandham, Shashidhar ;
Dawande, Milind ;
Prakash, Ravi .
JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2008, 68 (08) :1122-1134
[8]  
Husfeldt T., 2015, TOPICS CHROMATIC GRA, P277, DOI DOI 10.1017/CBO9781139519793.016
[9]  
Koloskova A, 2019, PR MACH LEARN RES, V97
[10]   Direct Acyclic Graph-Based Ledger for Internet of Things: Performance and Security Analysis [J].
Li, Yixin ;
Cao, Bin ;
Peng, Mugen ;
Zhang, Long ;
Zhang, Lei ;
Feng, Daquan ;
Yu, Jihong .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2020, 28 (04) :1643-1656