A Comprehensive Overview of Backdoor Attacks in Large Language Models Within Communication Networks

被引:6
作者
Yang, Haomiao [1 ,2 ]
Xiang, Kunlan [1 ,2 ]
Ge, Mengyu [3 ]
Li, Hongwei [1 ,2 ]
Lu, Rongxing [4 ]
Yu, Shui [5 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu 611731, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Cyber Secur, Chengdu 611731, Peoples R China
[3] ZTE Corp, RAN & Comp Power Syst Dept, Shenzhen 518055, Guangdong, Peoples R China
[4] Univ New Brunswick, Fac Comp Sci, Fredericton, NB E3B 5A3, Canada
[5] Univ Technol Sydney, Sch Comp Sci, Sydney, NSW 2007, Australia
来源
IEEE NETWORK | 2024年 / 38卷 / 06期
基金
中国国家自然科学基金;
关键词
Training; Computational modeling; Data models; Predictive models; Training data; Security; Solid modeling; Backdoor attacks; Large Language Models;
D O I
10.1109/MNET.2024.3367788
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The Large Language Models (LLMs) are poised to offer efficient and intelligent services for future mobile communication networks, owing to their exceptional capabilities in language comprehension and generation. However, the extremely high data and computational resource requirements for the performance of LLMs compel developers to resort to outsourcing training or utilizing third-party data and computing resources. These strategies may expose the model within the network to maliciously manipulated training data and processing, providing an opportunity for attackers to embed a hidden backdoor into the model, termed a backdoor attack. Backdoor attack in LLMs refers to embedding a hidden backdoor in LLMs that causes the model to perform normally on benign samples but exhibit degraded performance on poisoned ones. This issue is particularly concerning within communication networks where reliability and security are paramount. Despite the extensive research on backdoor attacks, there remains a lack of in-depth exploration specifically within the context of LLMs employed in communication networks, and a systematic review of such attacks is currently absent. In this survey, we systematically propose a taxonomy of backdoor attacks in LLMs as used in communication networks, dividing them into four major categories: input-triggered, prompt-triggered, instruction-triggered, and demonstration-triggered attacks. Furthermore, we conduct a comprehensive analysis of the benchmark datasets. Finally, we identify potential problems and open challenges, offering valuable insights into future research directions for enhancing the security and integrity of LLMs in communication networks.
引用
收藏
页码:211 / 218
页数:8
相关论文
共 15 条
[1]  
Bubeck S, 2023, Arxiv, DOI [arXiv:2303.12712, 10.48550/ARXIV.2303.12712]
[2]  
Cai Xiangrui, 2022, Advances in Neural Information Processing Systems
[3]  
Chen K., 2021, arXiv
[4]  
Chen LC, 2023, Arxiv, DOI arXiv:2305.02424
[5]  
Li L., 2021, arXiv
[6]   Hidden Backdoors in Human-Centric Language Models [J].
Li, Shaofeng ;
Liu, Hui ;
Dong, Tian ;
Zhao, Benjamin Zi Hao ;
Xue, Minhui ;
Zhu, Haojin ;
Lu, Jialiang .
CCS '21: PROCEEDINGS OF THE 2021 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2021, :3123-3140
[7]  
Lyu W., 2023, P ICLR WORKSH BACKD, P1
[8]  
Pan XD, 2022, PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, P3611
[9]  
Perez F, 2022, Arxiv, DOI arXiv:2211.09527
[10]  
Wang JX, 2023, Arxiv, DOI [arXiv:2305.14950, 10.48550/arXiv.2305.14950, DOI 10.48550/ARXIV.2305.14950]