Adversarial vulnerability bounds for Gaussian process classification

被引:3
作者
Smith, Michael Thomas [1 ]
Grosse, Kathrin [2 ]
Backes, Michael [3 ]
Alvarez, Mauricio A. [4 ]
机构
[1] Univ Sheffield, Dept Comp Sci, Sheffield, S Yorkshire, England
[2] Univ Cagliari, PRA Lab, Cagliari, Italy
[3] CISPA Helmholtz Ctr Informat Secur, Saarbrucken, Germany
[4] Univ Manchester, Dept Comp Sci, Manchester, Lancs, England
基金
英国工程与自然科学研究理事会;
关键词
Machine learning; Gaussian process; Adversarial example; Bound; Classification; Gaussian process classification;
D O I
10.1007/s10994-022-06224-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Protecting ML classifiers from adversarial examples is crucial. We propose that the main threat is an attacker perturbing a confidently classified input to produce a confident misclassification. We consider in this paper the L-0 attack in which a small number of inputs can be perturbed by the attacker at test-time. To quantify the risk of this form of attack we have devised a formal guarantee in the form of an adversarial bound (AB) for a binary, Gaussian process classifier using the EQ kernel. This bound holds for the entire input domain, bounding the potential of any future adversarial attack to cause a confident misclassification. We explore how to extend to other kernels and investigate how to maximise the bound by altering the classifier (for example by using sparse approximations). We test the bound using a variety of datasets and show that it produces relevant and practical bounds for many of them.
引用
收藏
页码:971 / 1009
页数:39
相关论文
共 38 条
[1]  
[Anonymous], 2017, arXiv
[2]  
[Anonymous], 2013, International Conference on Artificial Intelligence and Statistics
[3]  
Biggio B, 2013, JOINT EUROPEAN C MAC, DOI [10.1007/978-3-642-40994-3_25, DOI 10.1007/978-3-642-40994-3_25]
[4]  
Blaas A., 2020, 23 INT C ARTIFICIAL
[5]  
BOJCHEVSKI A, 2020, PR MACH LEARN RES, V119, pNIL96
[6]  
Cardelli L, 2019, AAAI CONF ARTIF INTE, P7759
[7]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[8]  
Carlini N, 2017, PROCEEDINGS OF THE 10TH ACM WORKSHOP ON ARTIFICIAL INTELLIGENCE AND SECURITY, AISEC 2017, P3, DOI 10.1145/3128572.3140444
[9]   Mode-finding for mixtures of Gaussian distributions [J].
Carreira-Perpiñán, MA .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (11) :1318-1323
[10]   Spam! [J].
Cranor, LF ;
LaMacchia, BA .
COMMUNICATIONS OF THE ACM, 1998, 41 (08) :74-83