Analysis of Privacy Leakage in Federated Large Language Models

被引:0
|
作者
Vu, Minh N. [1 ]
Nguyen, Truc [1 ]
Jeter, Tre' R. [1 ]
Thai, My T. [1 ]
机构
[1] Univ Florida, Gainesville, FL 32611 USA
来源
INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238 | 2024年 / 238卷
基金
美国国家科学基金会;
关键词
MEMBERSHIP INFERENCE ATTACKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid adoption of Federated Learning (FL) as the training and tuning protocol for applications utilizing Large Language Models (LLMs), recent research highlights the need for significant modifications to FL to accommodate the large-scale of LLMs. While substantial adjustments to the protocol have been introduced as a response, comprehensive privacy analysis for the adapted FL protocol is currently lacking. To address this gap, our work delves into an extensive examination of the privacy analysis of FL when used for training LLMs, both from theoretical and practical perspectives. In particular, we design two active membership inference attacks with guaranteed theoretical success rates to assess the privacy leakages of various adapted FL configurations. Our theoretical findings are translated into practical attacks, revealing substantial privacy vulnerabilities in popular LLMs, including BERT, RoBERTa, DistilBERT, and OpenAI's GPTs, across multiple real-world language datasets. Additionally, we conduct thorough experiments to evaluate the privacy leakage of these models when data is protected by state-of-the-art differential privacy (DP) mechanisms.
引用
收藏
页数:23
相关论文
共 50 条
  • [41] A Privacy Policy Text Compliance Reasoning Framework with Large Language Models for Healthcare Services
    Chen, Jintao
    Wang, Fan
    Pang, Shengye
    Chen, Mingshuai
    Xi, Meng
    Zhao, Tiancheng
    Yin, Jianwei
    TSINGHUA SCIENCE AND TECHNOLOGY, 2025, 30 (04): : 1831 - 1845
  • [42] Embedding Large Language Models into Extended Reality: Opportunities and Challenges for Inclusion, Engagement, and Privacy
    Bozkir, Efe
    Ozdel, Suleyman
    Lau, Ka Hei Carrie
    Wang, Mengdi
    Gao, Hong
    Kasneci, Enkelejda
    PROCEEDINGS OF THE 6TH CONFERENCE ON ACM CONVERSATIONAL USER INTERFACES, CUI 2024, 2024,
  • [43] Privacy-Preserving Techniques in Generative AI and Large Language Models: A Narrative Review
    Feretzakis, Georgios
    Papaspyridis, Konstantinos
    Gkoulalas-Divanis, Aris
    Verykios, Vassilios S.
    INFORMATION, 2024, 15 (11)
  • [44] Invited Paper: Security and Privacy in Large Language and Foundation Models: A Survey on GenAI Attacks
    Italiano, Giuseppe Francesco
    Martino, Alessandro
    Piccardo, Giorgio
    DISTRIBUTED COMPUTING AND INTELLIGENT TECHNOLOGY, ICDCIT 2025, 2025, 15507 : 1 - 17
  • [45] Privacy preserving large language models: ChatGPT case study based vision and framework
    Ullah, Imdad
    Hassan, Najm
    Gill, Sukhpal Singh
    Suleiman, Basem
    Ahanger, Tariq Ahamed
    Shah, Zawar
    Qadir, Junaid
    Kanhere, Salil S.
    IET Blockchain, 2024, 4 (S1): : 706 - 724
  • [46] PrivacyOracle: Configuring Sensor Privacy Firewalls with Large Language Models in Smart Built Environments
    Wang, Brian
    Garcia, Luis Antonio
    Srivastava, Mani
    PROCEEDINGS 45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS, SPW 2024, 2024, : 239 - 245
  • [47] InferDPT: Privacy-preserving Inference for Black-box Large Language Models
    Tong, Meng
    Chen, Kejiang
    Zhang, Jie
    Qi, Yuang
    Zhang, Weiming
    Yu, Nenghai
    Zhang, Tianwei
    Zhang, Zhikun
    arXiv, 2023,
  • [48] Towards Fair and Privacy-Preserving Federated Deep Models
    Lyu, Lingjuan
    Yu, Jiangshan
    Nandakumar, Karthik
    Li, Yitong
    Ma, Xingjun
    Jin, Jiong
    Yu, Han
    Ng, Kee Siong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2020, 31 (11) : 2524 - 2541
  • [49] Robust privacy amidst innovation with large language models through a critical assessment of the risks
    Chuang, Yao-Shun
    Sarkar, Atiquer Rahman
    Hsu, Yu-Chun
    Mohammed, Noman
    Jiang, Xiaoqian
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2025,
  • [50] dsSurvival: Privacy preserving survival models for federated individual patient meta-analysis in DataSHIELD
    Soumya Banerjee
    Ghislain N. Sofack
    Thodoris Papakonstantinou
    Demetris Avraam
    Paul Burton
    Daniela Zöller
    Tom R. P. Bishop
    BMC Research Notes, 15