Fedcs: Efficient communication scheduling in decentralized federated learning

被引:13
作者
Zong, Ruixing [1 ]
Qin, Yunchuan [1 ]
Wu, Fan [1 ]
Tang, Zhuo [1 ,2 ]
Li, Kenli [1 ,2 ]
机构
[1] Hunan Univ, Changsha, Peoples R China
[2] Natl Supercomp Changsha Ctr, Changsha, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Ring all reduce; Decentralized federated learning; Device placement; ATTACKS;
D O I
10.1016/j.inffus.2023.102028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decentralized federated learning is a training approach that prioritizes user data privacy protection, while also offering improved scalability and robustness. However, as the number of edge devices participating in training increases, a significant communication overhead arises among devices located in different geographical locations. Therefore, designing a well-thought-out gradient synchronization strategy is crucial for minimizing the overall communication overhead of training. To tackle this issue, this article introduces a 2D-Ring network structure based parameter synchronization strategy and a 2D-attention-based device placement algorithm, aiming to minimize communication overhead. The parameter synchronization strategy devises a two-layer circular communication architecture for the devices involved in training, thereby reducing the overall frequency of parameter synchronization in decentralized federated learning. By taking into account the total communication overhead and the device placement strategy, an optimization problem is formulated. Specifically, a 2D-attention neural network is constructed to optimize the device placement solution based on 2D-Ring network structure, leading to reduced communication overhead. Moreover, an evaluation model is designed to assess the communication overhead in a complex decentralized system during federated training. This enables precise determination of the total communication overhead throughout the training process, providing valuable insights for devising the device placement strategy. Extensive simulations confirm that the proposed approach achieves a substantial reductions of 55% and 64% in the total communication overhead for decentralized federated learning training with 50 and 100 devices, respectively.
引用
收藏
页数:9
相关论文
共 33 条
  • [1] A scalable, commodity data center network architecture
    Al-Fares, Mohammad
    Loukissas, Alexander
    Vahdat, Amin
    [J]. ACM SIGCOMM COMPUTER COMMUNICATION REVIEW, 2008, 38 (04) : 63 - 74
  • [2] An efficient concurrent implementation of a neural network algorithm
    Andonie, R.
    Chronopoulos, A. T.
    Grosu, D.
    Galmeanu, H.
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2006, 18 (12) : 1559 - 1573
  • [3] Beltran ETM, 2023, Arxiv, DOI arXiv:2211.08413
  • [4] Non-cooperative game algorithms for computation offloading in mobile edge computing environments
    Chen, Jianguo
    Deng, Qingying
    Yang, Xulei
    [J]. JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2023, 172 : 18 - 31
  • [5] Privacy-Preserving Deep Learning Model for Decentralized VANETs Using Fully Homomorphic Encryption and Blockchain
    Chen, Jianguo
    Li, Kenli
    Yu, Philip S.
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (08) : 11633 - 11642
  • [6] A Bi-layered Parallel Training Architecture for Large-Scale Convolutional Neural Networks
    Chen, Jianguo
    Li, Kenli
    Bilal, Kashif
    Zhou, Xu
    Li, Keqin
    Yu, Philip S.
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (05) : 965 - 976
  • [7] Chronopoulos A.T., 2002, A distributed discrete-time neural network architecture for pattern allocation and control
  • [8] Gabrielli E, 2023, Arxiv, DOI arXiv:2308.04604
  • [9] Hu CH, 2019, Arxiv, DOI arXiv:1908.07782
  • [10] Kingma D.P., 2014, arXiv, DOI [DOI 10.48550/ARXIV.1412.6980, 10.48550/arXiv.1412.6980]