Clean-label backdoor attack based on robust feature attenuation for speech recognition

被引:1
作者
Cai, Hanbo [1 ]
Zhang, Pengcheng [1 ]
Xiao, Yan [2 ]
Ji, Shunhui [1 ]
Xiao, Mingxuan [1 ]
Cheng, Letian [1 ]
机构
[1] Hohai Univ, Coll Comp Sci & Software Engn, Nanjing 211100, Jiangsu, Peoples R China
[2] Sun Yat sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
基金
中国国家自然科学基金;
关键词
Backdoor attack; Backdoor learning; Speech recognition; Trustworthy ML; MACHINE;
D O I
10.1016/j.eswa.2025.127546
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have been extensively and successfully applied in various speech recognition. Recent studies reveal that this type of model is susceptible to backdoor attacks, where adversaries can introduce malicious backdoors during the training phase of the victim's model. This paper focuses on poison-only backdoor attacks against speech recognition models. We identify that existing methods are not stealthy and perform inadequately in clean-label scenarios due to simplistic triggers, such as basic noise or conspicuous clips, which lack robust features. To address this, we propose the design of robust unnoticeable triggers by leveraging the structural features of the DNN model. Our approach involves attenuating robust features in the samples and creating triggers that integrate well with the DNN's structure, thus enhancing trigger effectiveness in clean-label situations. We ensure that our triggers are subtle by limiting their intensity and scope, thereby making our strategy stealthier. We have conducted extensive experiments on benchmark datasets, confirming our method's high efficacy in various scenarios, including fine-tuning, pruning, STRIP, spectral signatures, physical, and cross-model conditions, with an attack success rate of over 90% while remaining largely undetected. This research aims to raise awareness among researchers and developers about this potential threat and encourage the development of effective countermeasures. The code for reproducing main experiments is available at https://github.com/HanboCai/RFA-TUAP.
引用
收藏
页数:14
相关论文
共 102 条
[1]  
Nguyen A, 2015, PROC CVPR IEEE, P427, DOI 10.1109/CVPR.2015.7298640
[2]  
Athalye A, 2018, PR MACH LEARN RES, V80
[3]   Targeted Attack for Deep Hashing Based Retrieval [J].
Bai, Jiawang ;
Chen, Bin ;
Li, Yiming ;
Wu, Dongxian ;
Guo, Weiwei ;
Xia, Shu-Tao ;
Yang, En-Hui .
COMPUTER VISION - ECCV 2020, PT I, 2020, 12346 :618-634
[4]  
Barni M, 2019, IEEE IMAGE PROC, P101, DOI [10.1109/icip.2019.8802997, 10.1109/ICIP.2019.8802997]
[5]  
Beckert S, 2018, Columbia Stud Hist U, P1
[6]   Keyword Transformer: A Self-Attention Model for Keyword Spotting [J].
Berg, Axel ;
O'Connor, Mark ;
Cruz, Miguel Tairum .
INTERSPEECH 2021, 2021, :4249-4253
[7]   Toward Stealthy Backdoor Attacks Against Speech Recognition via Elements of Sound [J].
Cai, Hanbo ;
Zhang, Pengcheng ;
Dong, Hai ;
Xiao, Yan ;
Koffas, Stefanos ;
Li, Yiming .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 :5852-5866
[8]   Adversarial example-based test case generation for black-box speech recognition systems [J].
Cai, Hanbo ;
Zhang, Pengcheng ;
Dong, Hai ;
Grunske, Lars ;
Ji, Shunhui ;
Yuan, Tianhao .
SOFTWARE TESTING VERIFICATION & RELIABILITY, 2023, 33 (05)
[9]  
Chen K., 2022, ICLR
[10]  
Chen XY, 2017, Arxiv, DOI arXiv:1712.05526