Clean-label backdoor attack based on robust feature attenuation for speech recognition

被引:1
作者
Cai, Hanbo [1 ]
Zhang, Pengcheng [1 ]
Xiao, Yan [2 ]
Ji, Shunhui [1 ]
Xiao, Mingxuan [1 ]
Cheng, Letian [1 ]
机构
[1] Hohai Univ, Coll Comp Sci & Software Engn, Nanjing 211100, Jiangsu, Peoples R China
[2] Sun Yat sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
基金
中国国家自然科学基金;
关键词
Backdoor attack; Backdoor learning; Speech recognition; Trustworthy ML; MACHINE;
D O I
10.1016/j.eswa.2025.127546
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) have been extensively and successfully applied in various speech recognition. Recent studies reveal that this type of model is susceptible to backdoor attacks, where adversaries can introduce malicious backdoors during the training phase of the victim's model. This paper focuses on poison-only backdoor attacks against speech recognition models. We identify that existing methods are not stealthy and perform inadequately in clean-label scenarios due to simplistic triggers, such as basic noise or conspicuous clips, which lack robust features. To address this, we propose the design of robust unnoticeable triggers by leveraging the structural features of the DNN model. Our approach involves attenuating robust features in the samples and creating triggers that integrate well with the DNN's structure, thus enhancing trigger effectiveness in clean-label situations. We ensure that our triggers are subtle by limiting their intensity and scope, thereby making our strategy stealthier. We have conducted extensive experiments on benchmark datasets, confirming our method's high efficacy in various scenarios, including fine-tuning, pruning, STRIP, spectral signatures, physical, and cross-model conditions, with an attack success rate of over 90% while remaining largely undetected. This research aims to raise awareness among researchers and developers about this potential threat and encourage the development of effective countermeasures. The code for reproducing main experiments is available at https://github.com/HanboCai/RFA-TUAP.
引用
收藏
页数:14
相关论文
共 102 条
[71]  
Rouard S., 2021, IEEE INT C AC SPEECH, P21
[72]  
Saha A, 2020, AAAI CONF ARTIF INTE, V34, P11957
[73]  
Shafahi A, 2018, ADV NEUR IN, V31
[74]   Audio-domain Position-independent Backdoor Attack via Unnoticeable Triggers [J].
Shi, Cong ;
Zhang, Tianfang ;
Li, Zhuohang ;
Phan, Huy ;
Zhao, Tianming ;
Wang, Yan ;
Liu, Jian ;
Yuan, Bo ;
Chen, Yingying .
PROCEEDINGS OF THE 2022 THE 28TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, ACM MOBICOM 2022, 2022, :583-595
[75]  
Souri H., 2022, Advances in Neural Information Processing Systems
[76]   ATTENTION IS ALL YOU NEED IN SPEECH SEPARATION [J].
Subakan, Cem ;
Ravanelli, Mirco ;
Cornell, Samuele ;
Bronzi, Mirko ;
Zhong, Jianyuan .
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, :21-25
[77]  
Szegedy C, 2014, Arxiv, DOI arXiv:1312.6199
[78]  
Tang R., 2023, Advances in neural information processing systems (neurIPS), V36
[79]   Code Difference Guided Adversarial Example Generation for Deep Code Models [J].
Tian, Zhao ;
Chen, Junjie ;
Jin, Zhi .
2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, :850-862
[80]  
Tran B, 2018, ADV NEUR IN, V31