Training Provably Robust Models by Polyhedral Envelope Regularization

被引:6
作者
Liu, Chen [1 ]
Salzmann, Mathieu [1 ]
Susstrunk, Sabine [1 ]
机构
[1] Ecole Polytech Fed Lausanne EPFL, Sch Comp & Commun Sci, CH-1015 Lausanne, Switzerland
关键词
Robustness; Training; Predictive models; Computational modeling; Standards; Smoothing methods; Recurrent neural networks; Adversarial training; provable robustness;
D O I
10.1109/TNNLS.2021.3111892
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training certifiable neural networks enables us to obtain models with robustness guarantees against adversarial attacks. In this work, we introduce a framework to obtain a provable adversarial-free region in the neighborhood of the input data by a polyhedral envelope, which yields more fine-grained certified robustness than existing methods. We further introduce polyhedral envelope regularization (PER) to encourage larger adversarial-free regions and thus improve the provable robustness of the models. We demonstrate the flexibility and effectiveness of our framework on standard benchmarks; it applies to networks of different architectures and with general activation functions. Compared with state of the art, PER has negligible computational overhead; it achieves better robustness guarantees and accuracy on the clean data in various settings.
引用
收藏
页码:3146 / 3160
页数:15
相关论文
共 51 条
[1]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[2]  
Carmon Y, 2019, 33 C NEURAL INFORM P, V32
[3]  
Cheng M., 2019, 7 INT C LEARN REPR I, P1
[4]  
Cohen J, 2019, PR MACH LEARN RES, V97
[5]  
Croce F, 2020, PR MACH LEARN RES, V119
[6]  
Croce F, 2019, PR MACH LEARN RES, V89
[7]  
Dhillon GS, 2018, ARXIV PREPRINT ARXIV
[8]   Boosting Adversarial Attacks with Momentum [J].
Dong, Yinpeng ;
Liao, Fangzhou ;
Pang, Tianyu ;
Su, Hang ;
Zhu, Jun ;
Hu, Xiaolin ;
Li, Jianguo .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :9185-9193
[9]  
Du C., 2020, PROC INT C LEARN REP, P1, DOI DOI 10.48550/ARXIV.1905.10626
[10]   AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation [J].
Gehr, Timon ;
Mirman, Matthew ;
Drachsler-Cohen, Dana ;
Tsankov, Petar ;
Chaudhuri, Swarat ;
Vechev, Martin .
2018 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2018, :3-18