An adaptive Gaussian mixture method for nonlinear uncertainty propagation in neural networks

被引:10
作者
Zhang, Bin [1 ]
Shin, Yung C. [1 ]
机构
[1] Purdue Univ, Sch Mech Engn, W Lafayette, IN 47907 USA
关键词
Neural networks; Nonlinear uncertainty propagation; Gaussian mixture model; Dynamic systems; STATE;
D O I
10.1016/j.neucom.2021.06.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Using neural networks to address data-driven problems often entails dealing with uncertainties. However, the propagation of uncertainty through a network's nonlinear layers is usually a bottleneck, since the existing techniques designed to transmit Gaussian distributions via moment estimation are not capable of predicting non-Gaussian distributions. In this study, a Gaussian-mixture-based uncertainty propagation scheme is proposed for neural networks. Given that any input uncertainty can be characterized as a Gaussian mixture with a finite number of components, the developed scheme actively examines each mixture component and adaptively split those whose fidelity in representing uncertainty is deteriorated by the network's nonlinear activation layers. A Kullback-Leibler criterion that directly measures the nonlinearity-induced non-Gaussianity in post-activation distributions is derived to trigger splitting and a set of high-precision Gaussian splitting libraries is established. Four uncertainty propagation examples on dynamic systems and data-driven applications are demonstrated, in all of which the developed scheme exhibited exemplary fidelity and efficiency in predicting the evolution of non Gaussian distributions through both recurrent and multi-layer neural networks. (c) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页码:170 / 183
页数:14
相关论文
共 42 条
[11]   A method for state of energy estimation of lithium-ion batteries based on neural network model [J].
Dong, Guangzhong ;
Zhang, Xu ;
Zhang, Chenbin ;
Chen, Zonghai .
ENERGY, 2015, 90 :879-888
[12]   Hierarchical deep neural network for multivariate regression [J].
Du, Jun ;
Xu, Yong .
PATTERN RECOGNITION, 2017, 63 :149-157
[13]   The Split and Merge Unscented Gaussian Mixture Filter [J].
Faubel, Friedrich ;
McDonough, John ;
Klakow, Dietrich .
IEEE SIGNAL PROCESSING LETTERS, 2009, 16 (09) :786-789
[14]   Two moments of the logitnormal distribution [J].
Frederic, Patrizio ;
Lad, Frank .
COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2008, 37 (07) :1263-1269
[15]   Neural Network-Based Finite-Time Command Filtering Control for Switched Nonlinear Systems With Backlash-Like Hysteresis [J].
Fu, Cheng ;
Wang, Qing-Guo ;
Yu, Jinpeng ;
Lin, Chong .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (07) :3268-3273
[16]  
Gal Y, 2016, Uncertainty in deep learning
[17]   Lightweight Probabilistic Deep Networks [J].
Gast, Jochen ;
Roth, Stefan .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :3369-3378
[18]  
Golub G. H., 1996, Matrix computations, V3rd
[19]   Approximating the Kullback Leibler Divergence between Gaussian Mixture Models [J].
Hershey, John R. ;
Olsen, Peder A. .
2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL IV, PTS 1-3, 2007, :317-320
[20]   MULTILAYER FEEDFORWARD NETWORKS ARE UNIVERSAL APPROXIMATORS [J].
HORNIK, K ;
STINCHCOMBE, M ;
WHITE, H .
NEURAL NETWORKS, 1989, 2 (05) :359-366