FedITD: A Federated Parameter-Efficient Tuning With Pre-Trained Large Language Models and Transfer Learning Framework for Insider Threat Detection

被引:0
|
作者
Wang, Zhi Qiang [1 ]
Wang, Haopeng [1 ]
El Saddik, Abdulmotaleb [1 ,2 ]
机构
[1] Univ Ottawa, Sch Elect Engn & Comp Sci, Multimedia Commun Res Lab MCRLab, Ottawa, ON K1N 6N5, Canada
[2] MBZUAI, Dept Comp Vis, Abu Dhabi, U Arab Emirates
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Data models; Adaptation models; Threat assessment; Tuning; Security; Organizations; Costs; Computational modeling; Transfer learning; Deep learning; Computer security; Data augmentation; Artificial intelligence; Machine learning; Cybersecurity; insider threat; deep learning; transformer; BERT; RoBERTa; XLNet; DistilBERT; GPT; data augmentation; artificial intelligence; machine learning; pre-trained LLM; PETuning; adapter; LoRA; BitFit; LLM; NLP;
D O I
10.1109/ACCESS.2024.3482988
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Insider threats cause greater losses than external attacks, prompting organizations to invest in detection systems. However, there exist challenges: 1) Security and privacy concerns prevent data sharing, making it difficult to train robust models and identify new attacks. 2) The diversity and uniqueness of organizations require localized models, as a universal solution could be more effective. 3) High resource costs, delays, and data security concerns complicate building effective detection systems. This paper introduces FedITD, a flexible, hierarchy, and federated framework with local real-time detection systems, combining Large Language Models (LLM), Federated Learning (FL), Parameter Efficient Tuning (PETuning), and Transfer Learning (TF) for insider threat detection. FedITD uses FL to protect privacy while indirect integrating client information and employs PETuning methods (Adapter, BitFit, LoRA) with LLMs (BERT, RoBERTa, XLNet, DistilBERT) to reduce resource use and time delay. FedITD customizes client models and optimizes performance via transfer learning without central data transfer, further enhancing the detection of new attacks. FedITD outperforms other federated learning methods and its performance is very close to the best centrally trained method. Extensive experiment results show FedITD's superior performance, adaptability to varied data, and reduction of resource costs, achieving an optimal balance in detection capabilities across source data, unlabeled local data, and global data. Alternative PETuning implementations are also explored in this paper.
引用
收藏
页码:160396 / 160417
页数:22
相关论文
共 50 条
  • [1] FedPETuning: When Federated Learning Meets the Parameter-Efficient Tuning Methods of Pre-trained Language Models
    Zhang, Zhuo
    Yang, Yuanhang
    Dai, Yong
    Wang, Qifan
    Yu, Yue
    Que, Lizhen
    Xu, Zenglin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 9963 - 9977
  • [2] Parameter-efficient fine-tuning of large-scale pre-trained language models
    Ning Ding
    Yujia Qin
    Guang Yang
    Fuchao Wei
    Zonghan Yang
    Yusheng Su
    Shengding Hu
    Yulin Chen
    Chi-Min Chan
    Weize Chen
    Jing Yi
    Weilin Zhao
    Xiaozhi Wang
    Zhiyuan Liu
    Hai-Tao Zheng
    Jianfei Chen
    Yang Liu
    Jie Tang
    Juanzi Li
    Maosong Sun
    Nature Machine Intelligence, 2023, 5 : 220 - 235
  • [3] Parameter-efficient fine-tuning of large-scale pre-trained language models
    Ding, Ning
    Qin, Yujia
    Yang, Guang
    Wei, Fuchao
    Yang, Zonghan
    Su, Yusheng
    Hu, Shengding
    Chen, Yulin
    Chan, Chi-Min
    Chen, Weize
    Yi, Jing
    Zhao, Weilin
    Wang, Xiaozhi
    Liu, Zhiyuan
    Zheng, Hai-Tao
    Chen, Jianfei
    Liu, Yang
    Tang, Jie
    Li, Juanzi
    Sun, Maosong
    NATURE MACHINE INTELLIGENCE, 2023, 5 (03) : 220 - +
  • [4] REDUCING COMMUNICATION OVERHEAD IN FEDERATED LEARNING FOR PRE-TRAINED LANGUAGE MODELS USING PARAMETER-EFFICIENT FINETUNING
    Malaviya, Shubham
    Shukla, Manish
    Lodha, Sachin
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 456 - 469
  • [5] Parameter-Efficient Fine-Tuning of Pre-trained Large Language Models for Financial Text Analysis
    Langa, Kelly
    Wang, Hairong
    Okuboyejo, Olaperi
    ARTIFICIAL INTELLIGENCE RESEARCH, SACAIR 2024, 2025, 2326 : 3 - 20
  • [6] Neural Architecture Search for Parameter-Efficient Fine-tuning of Large Pre-trained Language Models
    Lawton, Neal
    Kumar, Anoop
    Thattai, Govind
    Galstyan, Aram
    Ver Steeg, Greg
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8506 - 8515
  • [7] ADT: An Additive Delta-Tuning approach for parameter-efficient tuning in pre-trained language models
    Li, Dong
    Tang, Jintao
    Li, Shasha
    Wang, Ting
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 382 - 386
  • [8] Hadamard Adapter: An Extreme Parameter-Efficient Adapter Tuning Method for Pre-trained Language Models
    Chen, Yuyan
    Fu, Qiang
    Fan, Ge
    Du, Lun
    Lou, Jian-Guang
    Han, Shi
    Zhang, Dongmei
    Li, Zhixu
    Xiao, Yanghua
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 276 - 285
  • [9] Federated Learning of Large Language Models with Parameter-Efficient Prompt Tuning and Adaptive Optimization
    Che, Tianshi
    Liu, Ji
    Zhou, Yang
    Ren, Jiaxiang
    Zhou, Jiwen
    Sheng, Victor S.
    Dai, Huaiyu
    Dou, Dejing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 7871 - 7888
  • [10] Bridging pre-trained models to continual learning: A hypernetwork based framework with parameter-efficient fine-tuning techniques
    Ding, Fengqian
    Xu, Chen
    Liu, Han
    Zhou, Bin
    Zhou, Hongchao
    INFORMATION SCIENCES, 2024, 674