A pruning feedforward small-world neural network by dynamic sparse regularization with smoothing l1/2 norm for nonlinear system modeling

被引:9
作者
Li, Wenjing [1 ]
Chu, Minghui
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Feedforward small-world neural network; Weight pruning; Dynamic regularization; Smoothingl1; 2; norm; CONVERGENCE; APPROXIMATION; ALGORITHM;
D O I
10.1016/j.asoc.2023.110133
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aiming to solve the problem of the relatively large architecture for the small-world neural network and improve its generalization ability, we propose a pruning feedforward small-world neural network based on a dynamic regularization method with the smoothing l1/2 norm (PFSWNN-DSRL1/2) and apply it to nonlinear system modeling. A feedforward small-world neural network is first constructed by the rewiring rule of Watts-Strogatz. By minimizing the modified error function added with a smoothing l1/2 norm, redundant weights are pruned to generate a sparse architecture. A dynamic adjusting strategy is further designed for the regularization strength to balance the tradeoff between the training accuracy and the sparsity. Several experiments are carried out to evaluate the performance of the proposed PFSWNN-DSRL1/2 on nonlinear system modeling. The results show that the PFSWNN-DSRL1/2 can achieve the satisfactory modeling accuracy with an average of 17% pruned weights. The comparative results demonstrate that the generalization performance of the proposed model is improved by 8.1% relative to the baseline method (FSWNN) but with a sparse structure, and the pruning does not degenerate its small-world property.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 43 条
[1]   A generalized feedforward neural network architecture for classification and regression [J].
Arulampalam, G ;
Bouzerdoum, A .
NEURAL NETWORKS, 2003, 16 (5-6) :561-568
[2]   Pruning algorithms of neural networks - a comparative study [J].
Augasta, M. Gethsiyal ;
Kathirvalavakumar, T. .
OPEN COMPUTER SCIENCE, 2013, 3 (03) :105-115
[3]   Complex brain networks: graph theoretical analysis of structural and functional systems [J].
Bullmore, Edward T. ;
Sporns, Olaf .
NATURE REVIEWS NEUROSCIENCE, 2009, 10 (03) :186-198
[4]   Construction of feedforward neural networks with simple architectures and approximation abilities [J].
Chen, Zhixiang ;
Cao, Feilong .
MATHEMATICAL METHODS IN THE APPLIED SCIENCES, 2021, 44 (02) :1788-1795
[5]   Communication on the structure of biological networks [J].
Deyasi, Krishanu ;
Upadhyay, Shashankaditya ;
Banerjee, Anirban .
PRAMANA-JOURNAL OF PHYSICS, 2016, 86 (03) :617-635
[6]   Impact of small-world network topology on the conventional artificial neural network for the diagnosis of diabetes [J].
Erkaymaz, Okan ;
Ozer, Mahmut .
CHAOS SOLITONS & FRACTALS, 2016, 83 :178-185
[7]   Performance Analysis of A Feed-Forward Artifical Neural Network With Small-World Topology [J].
Erkaymaz, Okan ;
Ozer, Mahmut ;
Yumusak, Nejat .
FIRST WORLD CONFERENCE ON INNOVATION AND COMPUTER SCIENCES (INSODE 2011), 2012, 1 :291-296
[8]   Modeling of Batch Processes Using Explicitly Time-Dependent Artificial Neural Networks [J].
Ganesh, Botla ;
Kumar, Vadlagattu Varun ;
Rani, Kalipatnapu Yamuna .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2014, 25 (05) :970-979
[9]   Using adjacency matrices to lay out larger small-world networks [J].
Gibson, Helen ;
Vickers, Paul .
APPLIED SOFT COMPUTING, 2016, 42 :80-92
[10]   Regularization Techniques and Suboptimal Solutions to Optimization Problems in Learning from Data [J].
Gnecco, Giorgio ;
Sanguineti, Marcello .
NEURAL COMPUTATION, 2010, 22 (03) :793-829