Training Provably Robust Models by Polyhedral Envelope Regularization

被引:6
作者
Liu, Chen [1 ]
Salzmann, Mathieu [1 ]
Susstrunk, Sabine [1 ]
机构
[1] Ecole Polytech Fed Lausanne EPFL, Sch Comp & Commun Sci, CH-1015 Lausanne, Switzerland
关键词
Robustness; Training; Predictive models; Computational modeling; Standards; Smoothing methods; Recurrent neural networks; Adversarial training; provable robustness;
D O I
10.1109/TNNLS.2021.3111892
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training certifiable neural networks enables us to obtain models with robustness guarantees against adversarial attacks. In this work, we introduce a framework to obtain a provable adversarial-free region in the neighborhood of the input data by a polyhedral envelope, which yields more fine-grained certified robustness than existing methods. We further introduce polyhedral envelope regularization (PER) to encourage larger adversarial-free regions and thus improve the provable robustness of the models. We demonstrate the flexibility and effectiveness of our framework on standard benchmarks; it applies to networks of different architectures and with general activation functions. Compared with state of the art, PER has negligible computational overhead; it achieves better robustness guarantees and accuracy on the clean data in various settings.
引用
收藏
页码:3146 / 3160
页数:15
相关论文
共 51 条
[51]  
Zhang H, 2018, ADV NEUR IN, V31