A Computationally Efficient Weight Pruning Algorithm for Artificial Neural Network Classifiers

被引:0
作者
Sakshi [1 ]
Kumar, Ravi [1 ]
机构
[1] Thapar Univ, Elect & Commun Engn Dept, Patiala 147004, Punjab, India
关键词
Weight pruning; Artificial neural network; Backpropagation; Complexity penalty; Fisher information; Pattern classification; MULTILAYER PERCEPTRONS;
D O I
10.1007/s13369-017-2887-2
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
A novel technique is being proposed to prune the weights of artificial neural networks (ANNs) while training with backpropagation algorithm. Iterative update of weights through gradient descent mechanism does not guarantee convergence in a specified number of epochs. Pruning of non-relevant weights not only reduces the computational complexity but also improves the classification performance. This algorithm first defines the relevance of initialized weights in a statistical sense by introducing a coefficient of dominance for each weight converging on a hidden node and subsequently employing the concept of complexity penalty. Based upon complexity penalty for each weight, a decision has been taken to either prune or retain the weight. It has been shown analytically that a weight with higher complexity penalty has a high degree of Fisher information which further implies its ability to capture the variations in the input set for better classification. Simulation experiments performed with five benchmark data sets reveal that ANNs trained after being pruned using the proposed technique exhibit higher convergence, lower execution time and higher success rate in the test phase and yields substantial reduction in computational resources. For complex architectures, early convergence was found to be directly correlated with percentage of weights pruned. The efficacy of the technique has been validated on several benchmark datasets having large diversity of attributes.
引用
收藏
页码:6787 / 6799
页数:13
相关论文
共 33 条
  • [21] Constructive neural-network learning algorithms for pattern classification
    Parekh, R
    Yang, JH
    Honavar, V
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2000, 11 (02): : 436 - 451
  • [22] Going Deeper with Embedded FPGA Platform for Convolutional Neural Network
    Qiu, Jiantao
    Wang, Jie
    Yao, Song
    Guo, Kaiyuan
    Li, Boxun
    Zhou, Erjin
    Yu, Jincheng
    Tang, Tianqi
    Xu, Ningyi
    Song, Sen
    Wang, Yu
    Yang, Huazhong
    [J]. PROCEEDINGS OF THE 2016 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS (FPGA'16), 2016, : 26 - 35
  • [23] PRUNING ALGORITHMS - A SURVEY
    REED, R
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1993, 4 (05): : 740 - 747
  • [24] ADVANCED SUPERVISED LEARNING IN MULTILAYER PERCEPTRONS - FROM BACKPROPAGATION TO ADAPTIVE LEARNING ALGORITHMS
    RIEDMILLER, M
    [J]. COMPUTER STANDARDS & INTERFACES, 1994, 16 (03) : 265 - 278
  • [25] Sabo D., IEEE INT JOINT C NEU, P3313
  • [26] ISAAC: A Convolutional Neural Network Accelerator with In-Situ Analog Arithmetic in Crossbars
    Shafiee, Ali
    Nag, Anirban
    Muralimanohar, Naveen
    Balasubramonian, Rajeev
    Strachan, John Paul
    Hu, Miao
    Williams, R. Stanley
    Srikumar, Vivek
    [J]. 2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, : 14 - 26
  • [27] Exploring neuro-genetic processing of electronic nose data
    Srivastava, AK
    Shukla, KK
    Srivastava, SK
    [J]. MICROELECTRONICS JOURNAL, 1998, 29 (11) : 921 - 931
  • [28] Multilayer perceptrons: Approximation order and necessary number of hidden units
    Trenn, Stephan
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (05): : 836 - 844
  • [29] Enhancing the generalization ability of neural networks through controlling the hidden layers
    Wan, Weishui
    Mabu, Shingo
    Shimada, Kaoru
    Hirasawa, Kotaro
    Hu, Jinglu
    [J]. APPLIED SOFT COMPUTING, 2009, 9 (01) : 404 - 414
  • [30] Weigend AS, 1990, ADV NEURAL INFORM PR, P875