Federated Large Language Model: Solutions, Challenges and Future Directions

被引:0
作者
Hu, Jiahui [1 ]
Wang, Dan [2 ]
Wang, Zhibo
Pang, Xiaoyi [1 ]
Xu, Huiyu [1 ]
Ren, Ju [3 ,4 ]
Ren, Kui [1 ]
机构
[1] Zhejiang Univ, State Key Lab Blockchain & Data Secur, Hangzhou, Peoples R China
[2] Hunan Normal Univ, Informat Sci & Engn, Changsha, Peoples R China
[3] Tsinghua Univ, Tsinghua, Peoples R China
[4] Zhongguancun Lab, Zhongguancun, Peoples R China
基金
中国博士后科学基金; 国家重点研发计划; 中国国家自然科学基金;
关键词
Training; LoRa; Servers; Computational modeling; Data models; Tuning; Transformers; Adaptation models; Large language models; Privacy;
D O I
10.1109/MWC.009.2400244
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Large language models (LLMs) have become increasingly popular due to their exceptional performance in various artificial intelligence applications. However, their development often suffers from the scarcity of high-quality data and the extensive requirements for computing resources. These obstacles are even more severe for enterprises in vertical industries, which have limited computer resources but urgently require large-scale models for specific activities. To address these issues, LLMs call for the integration of federated learning (FL), which enables the collaborative learning of a powerful LLM using private data and computing resources from multiple entities. In this article, we present a systematic introduction to the federated large language model (Fed-LLM), a distributed learning of LLM in the FL manner. We first introduce the learning paradigm of Fed-LLM, which is called federated parameter-efficient fine-tuning (Fed-PEFT). Fed-PEFT empowers the collaborative fine-tuning of pre-trained LLMs by only involving a small subset of parameters in local LLMs. Specifically, we detail the workflow of Fed-PEFT, and summarize the state-of-the-art solutions in this area. Additionally, we discuss the challenges faced in Fed-LLMs, including efficiency, privacy, and security. Finally, we introduce future directions to facilitate the research of Fed-LLMs and guide coming explorations in this nascent field.
引用
收藏
页数:8
相关论文
共 15 条
  • [1] Babakniya S, 2023, Arxiv, DOI arXiv:2308.06522
  • [2] Cai DQ, 2022, Arxiv, DOI arXiv:2205.10162
  • [3] Che TS, 2024, Arxiv, DOI arXiv:2310.15080
  • [4] Cho YJ, 2024, Arxiv, DOI arXiv:2401.06432
  • [5] Guo T., 2023, P ACM WEB C 2023, P1364
  • [6] Houlsby N, 2019, PR MACH LEARN RES, V97
  • [7] Hu EJ, 2021, Arxiv, DOI arXiv:2106.09685
  • [8] Kim Y., 2023, FINDINGS ASS COMPUTA, P1159
  • [9] Liao BH, 2023, Arxiv, DOI arXiv:2305.16742
  • [10] Li XL, 2021, Arxiv, DOI arXiv:2101.00190