Audio-domain Position-independent Backdoor Attack via Unnoticeable Triggers

被引:30
作者
Shi, Cong [1 ]
Zhang, Tianfang [1 ]
Li, Zhuohang [2 ]
Phan, Huy [1 ]
Zhao, Tianming [3 ]
Wang, Yan [3 ]
Liu, Jian [2 ]
Yuan, Bo [1 ]
Chen, Yingying [1 ]
机构
[1] Rutgers State Univ, Newark, NJ 07102 USA
[2] Univ Tennessee, Knoxville, TN USA
[3] Temple Univ, Philadelphia, PA USA
来源
PROCEEDINGS OF THE 2022 THE 28TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, ACM MOBICOM 2022 | 2022年
基金
美国国家科学基金会;
关键词
Audio-domain Backdoor Attacks; Position-independent Attacks; Over-the-air Physical Attacks;
D O I
10.1145/3495243.3560531
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Deep learning models have become key enablers of voice user interfaces. With the growing trend of adopting outsourced training of these models, backdoor attacks, stealthy yet effective training-phase attacks, have gained increasing attention. They inject hidden trigger patterns through training set poisoning and overwrite the model's predictions in the inference phase. Research in backdoor attacks has been focusing on image classification tasks, while there have been few studies in the audio domain. In this work, we explore the severity of audio-domain backdoor attacks and demonstrate their feasibility under practical scenarios of voice user interfaces, where an adversary injects (plays) an unnoticeable audio trigger into live speech to launch the attack. To realize such attacks, we consider jointly optimizing the audio trigger and the target model in the training phase, deriving a position-independent, unnoticeable, and robust audio trigger. We design new data poisoning techniques and penalty-based algorithms that inject the trigger into randomly generated temporal positions in the audio input during training, rendering the trigger resilient to any temporal position variations. We further design an environmental sound mimicking technique to make the trigger resemble unnoticeable situational sounds and simulate played over-the-air distortions to improve the trigger's robustness during the joint optimization process. Extensive experiments on two important applications (i.e., speech command recognition and speaker recognition) demonstrate that our attack can achieve an average success rate of over 99% under both digital and physical attack settings.
引用
收藏
页码:583 / 595
页数:13
相关论文
共 45 条
[1]  
Aghakhani H., 2020, Computing Research Repository (CoRR)
[2]   IMAGE METHOD FOR EFFICIENTLY SIMULATING SMALL-ROOM ACOUSTICS [J].
ALLEN, JB ;
BERKLEY, DA .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1979, 65 (04) :943-950
[3]  
Amazon, 2022, Amazon SageMaker
[4]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[5]   Audio Adversarial Examples: Targeted Attacks on Speech-to-Text [J].
Carlini, Nicholas ;
Wagner, David .
2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY WORKSHOPS (SPW 2018), 2018, :1-7
[6]   Who is Real Bob? Adversarial Attacks on Speaker Recognition Systems [J].
Chen, Guangke ;
Chen, Sen ;
Fan, Lingling ;
Du, Xiaoning ;
Zhao, Zhe ;
Song, Fu ;
Liu, Yang .
2021 IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP, 2021, :694-711
[7]  
Chen XY, 2017, Arxiv, DOI arXiv:1712.05526
[8]  
de Andrade DC, 2018, Arxiv, DOI arXiv:1808.08929
[9]   A tutorial on the cross-entropy method [J].
De Boer, PT ;
Kroese, DP ;
Mannor, S ;
Rubinstein, RY .
ANNALS OF OPERATIONS RESEARCH, 2005, 134 (01) :19-67
[10]   STRIP: A Defence Against Trojan Attacks on Deep Neural Networks [J].
Gao, Yansong ;
Xu, Change ;
Wang, Derui ;
Chen, Shiping ;
Ranasinghe, Damith C. ;
Nepal, Surya .
35TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSA), 2019, :113-125