A Comparative Study of Sparsity Promoting Techniques in Neural Network for Modeling Non-Linear Dynamics

被引:0
作者
Haugstvedt, Emil Johannesen [1 ]
Calero, Alberto Mino [1 ]
Lundby, Erlend Torje Berg [1 ]
Rasheed, Adil [1 ]
Gravdahl, Jan Tommy [1 ]
机构
[1] Norwegian Univ Sci & Technol, Dept Engn Cybernet, N-7491 Trondheim, Norway
关键词
Neural networks; Dynamical systems; Mathematical models; Biological neural networks; Differential equations; Carbon; Aluminum; Electrochemical processes; Nonlinear dynamical systems; Aluminum electrolysis; data-driven modeling; nonlinear dynamics; ordinary differential equations; sparse neural networks;
D O I
10.1109/ACCESS.2023.3336292
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sparsity-promoting techniques show promising results in improving the generalization of neural networks. However, the literature contains limited information on how different sparsity techniques affect generalization when using neural networks to model non-linear dynamical systems. This study examines the use of sparsity-enhancing techniques to improve accuracy and reduce the divergence rate of neural networks used to simulate such systems. A range of sparsity methods, including hard and soft thresholding, pruning and regrowing, and L1-regularization, were applied to neural networks and evaluated in a complex nonlinear aluminum extraction process by electrolysis. The results showed that the most effective technique was L1 regularization, which enhanced the important connections in the network and improved the model performance. In contrast, many of the more advanced sparsity techniques resulted in significantly worse performance and higher divergence rates. Additionally, the application of Stochastic Weight Averaging during training increased performance and reduced the number of diverging simulations. These findings suggest that carefully selecting the right sparsity techniques and model structures can improve the performance of neural network-based simulations of dynamical systems.
引用
收藏
页码:131435 / 131452
页数:18
相关论文
共 47 条
[1]   Effective gene expression prediction from sequence by integrating long-range interactions [J].
Avsec, Ziga ;
Agarwal, Vikram ;
Visentin, Daniel ;
Ledsam, Joseph R. ;
Grabska-Barwinska, Agnieszka ;
Taylor, Kyle R. ;
Assael, Yannis ;
Jumper, John ;
Kohli, Pushmeet ;
Kelley, David R. .
NATURE METHODS, 2021, 18 (10) :1196-+
[2]  
Binous H., Duffing Oscillator
[3]  
Cangelosi R., Lotka-Volterra Competition Model
[4]  
Chauvin Yves, 1988, ADV NEURAL INFORM PR, V1
[5]   Review of Image Classification Algorithms Based on Convolutional Neural Networks [J].
Chen, Leiyu ;
Li, Shaobo ;
Bai, Qiang ;
Yang, Jing ;
Jiang, Sanlong ;
Miao, Yanming .
REMOTE SENSING, 2021, 13 (22)
[6]   DE-NOISING BY SOFT-THRESHOLDING [J].
DONOHO, DL .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1995, 41 (03) :613-627
[7]   Semi-empirical Neural Network Based Approach to Modelling and Simulation of Controlled Dynamical Systems [J].
Egorchev, Mikhail, V ;
Tiumentsev, Yury, V .
8TH ANNUAL INTERNATIONAL CONFERENCE ON BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, BICA 2017 (EIGHTH ANNUAL MEETING OF THE BICA SOCIETY), 2018, 123 :134-139
[8]  
Gallicchio C, 2020, AAAI CONF ARTIF INTE, V34, P3898
[9]  
Garipov T., 2018, arXiv
[10]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1