Personalized federated learning based on multi-head attention algorithm

被引:0
作者
Shanshan Jiang
Meixia Lu
Kai Hu
Jiasheng Wu
Yaogen Li
Liguo Weng
Min Xia
Haifeng Lin
机构
[1] Nanjing University of Information Science and Technology,School of management Science and Engineering
[2] Nanjing University of Information Science and Technology,Jiangsu Provincial Collaborative Innovation Center for Atmospheric Environment and Equipment Technology
[3] Nanjing Forestry University,College of Information Science and Technology
来源
International Journal of Machine Learning and Cybernetics | 2023年 / 14卷
关键词
Federated learning; Multi-head attention mechanism; Personalize;
D O I
暂无
中图分类号
学科分类号
摘要
Federated Learning (FL) is an algorithm for the encrypted exchange of model parameters while ensuring the independence of participants. Classic federated learning does not take into account the correlation between features, nor does it take into account the data differences caused by the reasonable personalization of each client. Therefore, this paper proposes a personalized federated learning algorithm based on a multi-head attention mechanism. First, in order to improve the personalization of local models, attention mechanism is used to capture the relevance of local features. Then, when aggregating local models, the weight λ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda$$\end{document} is generated for local models based on the differences between models, and finally aggregate them into a new global model. Finally, the multi-head attention is proposed to calculate the importance score of the global model parameters on the current local model, and assign it to the local model as the attention coefficient, so as to realize personalized federated learning. Through experiments on MNIST, SVHN and STL10 datasets, the validity of Personalized Federated Learning is verified, and the rationality of hyperparameter setting is discussed through visualizing results.
引用
收藏
页码:3783 / 3798
页数:15
相关论文
共 93 条
  • [1] Honghao Gao(2022)Ppo2: Location privacy-oriented task offloading to edge computing using reinforcement learning for intelligent autonomous transport systems IEEE Trans Transp Syst 17 15-1282
  • [2] Wanqiu Huang(2021)A weakly supervised semantic segmentation network by aggregating seed cues: the multi-object proposal generation perspective ACM Trans Multimed Comput Commun Appl 54 1273-3157
  • [3] Tong Liu(2022)Tsmae: A novel anomaly detection approach for internet of things time series data using memory-augmented autoencoder IEEE Trans Netw Sci Eng 30 3149-6162
  • [4] Yuyu Yin(2016)Communication-efficient learning of deep networks from decentralized data PMLR 34 6149-5980
  • [5] Youhuizi Li(2017)A highly efficient gradient boosting decision tree NIPS 33 5973-345
  • [6] Xiao Junsheng(2022)Multi-scale strip pooling feature aggregation network for cloud and cloud shadow segmentation Neural Comput Appl 57 336-98
  • [7] Huahu Xu(2019)Dan: deep attention neural network for news recommendation AAAI 36 7-19
  • [8] Gao Honghao(2019)Neural news recommendation with long-and short-term user representations ACL 10 1-760
  • [9] Bian Minjie(2019)and Qiang Yang. SecureBoost: A lossless federated learning framework IEEE Intell Sys 51 751-2324
  • [10] Li Yang(2019)Federated machine learning: concept and applications ACM 86 2278-43