Robust deep neural network surrogate models with uncertainty quantification via adversarial training

被引:0
作者
Zhang, Lixiang [1 ]
Li, Jia [1 ]
机构
[1] Penn State Univ, Dept Stat, University Pk, PA 16802 USA
基金
美国国家科学基金会;
关键词
adversarial training; robustness; simulator; surrogate model; uncertainty quantification; GAUSSIAN PROCESS;
D O I
10.1002/sam.11610
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Surrogate models have been used to emulate mathematical simulators of physical or biological processes for computational efficiency. High-speed simulation is crucial for conducting uncertainty quantification (UQ) when the simulation must repeat over many randomly sampled input points (aka the Monte Carlo method). A simulator can be so computationally intensive that UQ is only feasible with a surrogate model. Recently, deep neural network (DNN) surrogate models have gained popularity for their state-of-the-art emulation accuracy. However, it is well-known that DNN is prone to severe errors when input data are perturbed in particular ways, the very phenomenon which has inspired great interest in adversarial training. In the case of surrogate models, the concern is less about a deliberate attack exploiting the vulnerability of a DNN but more of the high sensitivity of its accuracy to input directions, an issue largely ignored by researchers using emulation models. In this paper, we show the severity of this issue through empirical studies and hypothesis testing. Furthermore, we adopt methods in adversarial training to enhance the robustness of DNN surrogate models. Experiments demonstrate that our approaches significantly improve the robustness of the surrogate models without compromising emulation accuracy.
引用
收藏
页码:295 / 304
页数:10
相关论文
共 38 条
[1]   Bayesian History Matching of Complex Infectious Disease Models Using Emulation: A Tutorial and a Case Study on HIV in Uganda [J].
Andrianakis, Ioannis ;
Vernon, Ian R. ;
McCreesh, Nicky ;
McKinley, Trevelyan J. ;
Oakley, Jeremy E. ;
Nsubuga, Rebecca N. ;
Goldstein, Michael ;
White, Richard G. .
PLOS COMPUTATIONAL BIOLOGY, 2015, 11 (01)
[2]  
Balakrishnama S., 1998, Linear discriminant analysis‐a brief tutorial, V18, P1
[3]   Multi-output separable Gaussian process: Towards an efficient, fully Bayesian paradigm for uncertainty quantification [J].
Bilionis, Ilias ;
Zabaras, Nicholas ;
Konomi, Bledar A. ;
Lin, Guang .
JOURNAL OF COMPUTATIONAL PHYSICS, 2013, 241 :212-239
[4]   Multi-output local Gaussian process regression: Applications to uncertainty quantification [J].
Bilionis, Ilias ;
Zabaras, Nicholas .
JOURNAL OF COMPUTATIONAL PHYSICS, 2012, 231 (17) :5718-5746
[5]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[6]  
Chakraborty A, 2018, Arxiv, DOI arXiv:1810.00069
[7]  
Cheng YY, 2022, Arxiv, DOI arXiv:2110.12734
[8]   Inverting the Generator of a Generative Adversarial Network [J].
Creswell, Antonia ;
Bharath, Anil Anthony .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (07) :1967-1974
[9]   Denoising Adversarial Autoencoders [J].
Creswell, Antonia ;
Bharath, Anil Anthony .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2019, 30 (04) :968-984
[10]   Generative Adversarial Networks An overview [J].
Creswell, Antonia ;
White, Tom ;
Dumoulin, Vincent ;
Arulkumaran, Kai ;
Sengupta, Biswa ;
Bharath, Anil A. .
IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (01) :53-65