A pruning feedforward small-world neural network by dynamic sparse regularization with smoothing l1/2 norm for nonlinear system modeling

被引:9
作者
Li, Wenjing [1 ]
Chu, Minghui
机构
[1] Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Feedforward small-world neural network; Weight pruning; Dynamic regularization; Smoothingl1; 2; norm; CONVERGENCE; APPROXIMATION; ALGORITHM;
D O I
10.1016/j.asoc.2023.110133
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Aiming to solve the problem of the relatively large architecture for the small-world neural network and improve its generalization ability, we propose a pruning feedforward small-world neural network based on a dynamic regularization method with the smoothing l1/2 norm (PFSWNN-DSRL1/2) and apply it to nonlinear system modeling. A feedforward small-world neural network is first constructed by the rewiring rule of Watts-Strogatz. By minimizing the modified error function added with a smoothing l1/2 norm, redundant weights are pruned to generate a sparse architecture. A dynamic adjusting strategy is further designed for the regularization strength to balance the tradeoff between the training accuracy and the sparsity. Several experiments are carried out to evaluate the performance of the proposed PFSWNN-DSRL1/2 on nonlinear system modeling. The results show that the PFSWNN-DSRL1/2 can achieve the satisfactory modeling accuracy with an average of 17% pruned weights. The comparative results demonstrate that the generalization performance of the proposed model is improved by 8.1% relative to the baseline method (FSWNN) but with a sparse structure, and the pruning does not degenerate its small-world property.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:14
相关论文
共 43 条
[11]   On the approximation by single hidden layer feedforward neural networks with fixed weights [J].
Guliyev, Namig J. ;
Ismailov, Vugar E. .
NEURAL NETWORKS, 2018, 98 :296-304
[12]   Research on Trim of Multilayer Feedforward Small World Network Based on E-exponential Information Entropy [J].
Guo, Dongwei ;
Liu, Yang .
2017 NINTH INTERNATIONAL CONFERENCE ON INTELLIGENT HUMAN-MACHINE SYSTEMS AND CYBERNETICS (IHMSC 2017), VOL 2, 2017, :155-158
[13]  
KROGH A, 1992, ADV NEUR IN, V4, P950
[14]   A pruning feedforward small-world neural network based on Katz centrality for nonlinear system modeling [J].
Li, Wenjing ;
Chu, Minghui ;
Qiao, Junfei .
NEURAL NETWORKS, 2020, 130 :269-285
[15]  
Li X, 2011, 2011 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTATION AND INDUSTRIAL APPLICATION (ICIA2011), VOL I, P353, DOI 10.1109/CSAE.2011.5952696
[16]   Sparse logistic regression with a L1/2 penalty for gene selection in cancer classification [J].
Liang, Yong ;
Liu, Cheng ;
Luan, Xin-Ze ;
Leung, Kwong-Sak ;
Chan, Tak-Ming ;
Xu, Zong-Ben ;
Zhang, Hai .
BMC BIOINFORMATICS, 2013, 14
[17]   Smoothed L1/2 regularizer learning for split-complex valued neuro-fuzzy algorithm for TSK system and its convergence results [J].
Liu, Yan ;
Yang, Dakun ;
Li, Feng .
JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2018, 355 (13) :6132-6151
[18]   Convergence of batch gradient learning algorithm with smoothing L1/2 regularization for Sigma-Pi-Sigma neural networks [J].
Liu, Yan ;
Li, Zhengxue ;
Yang, Dakun ;
Mohamed, Kh. Sh. ;
Wang, Jing ;
Wu, Wei .
NEUROCOMPUTING, 2015, 151 :333-341
[19]   Robust nonlinear system identification using neural-network models [J].
Lu, SW ;
Basar, T .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (03) :407-429
[20]   Regression and classification using extreme learning machine based on L1-norm and L2-norm [J].
Luo, Xiong ;
Chang, Xiaohui ;
Ban, Xiaojuan .
NEUROCOMPUTING, 2016, 174 :179-186