Structured pruning of recurrent neural networks through neuron selection

被引:30
作者
Wen, Liangjian [1 ]
Zhang, Xuanyang [1 ]
Bai, Haoli [2 ]
Xu, Zenglin [1 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, SMILE Lab, Chengdu 610031, Peoples R China
[2] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong 999077, Peoples R China
[3] Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
关键词
Feature selection; Recurrent neural networks; Learning sparse models; Model compression;
D O I
10.1016/j.neunet.2019.11.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) have recently achieved remarkable successes in a number of applications. However, the huge sizes and computational burden of these models make it difficult for their deployment on edge devices. A practically effective approach is to reduce the overall storage and computation costs of RNNs by network pruning techniques. Despite their successful applications, those pruning methods based on Lasso either produce irregular sparse patterns in weight matrices, which is not helpful in practical speedup. To address these issues, we propose a structured pruning method through neuron selection which can remove the independent neuron of RNNs. More specifically, we introduce two sets of binary random variables, which can be interpreted as gates or switches to the input neurons and the hidden neurons, respectively. We demonstrate that the corresponding optimization problem can be addressed by minimizing the L-0 norm of the weight matrix. Finally, experimental results on language modeling and machine reading comprehension tasks have indicated the advantages of the proposed method in comparison with state-of-the-art pruning competitors. In particular, nearly 20x practical speedup during inference was achieved without losing performance for the language model on the Penn TreeBank dataset, indicating the promising performance of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:134 / 141
页数:8
相关论文
共 50 条
  • [41] Pseudo Random Number Generation through Reinforcement Learning and Recurrent Neural Networks
    Pasqualini, Luca
    Parton, Maurizio
    ALGORITHMS, 2020, 13 (11)
  • [42] Structured Network Pruning via Adversarial Multi-indicator Architecture Selection
    Yuxuan Wei
    Ying Chen
    Circuits, Systems, and Signal Processing, 2021, 40 : 4127 - 4143
  • [43] Structured Network Pruning via Adversarial Multi-indicator Architecture Selection
    Wei, Yuxuan
    Chen, Ying
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 2021, 40 (08) : 4127 - 4143
  • [44] Application of Convolutional Neural Networks and Recurrent Neural Networks in Food Safety
    Ding, Haohan
    Hou, Haoke
    Wang, Long
    Cui, Xiaohui
    Yu, Wei
    Wilson, David I.
    FOODS, 2025, 14 (02)
  • [45] Feature Selection With Neural Networks
    Philippe Leray
    Patrick Gallinari
    Behaviormetrika, 1999, 26 (1) : 145 - 166
  • [46] Feature selection with neural networks
    Verikas, A
    Bacauskiene, M
    PATTERN RECOGNITION LETTERS, 2002, 23 (11) : 1323 - 1335
  • [47] Dynamic connection pruning for densely connected convolutional neural networks
    Xinyi Hu
    Hangxiang Fang
    Ling Zhang
    Xue Zhang
    Howard H. Yang
    Dongxiao Yang
    Bo Peng
    Zheyang Li
    Haoji Hu
    Applied Intelligence, 2023, 53 : 19505 - 19521
  • [48] Dynamic connection pruning for densely connected convolutional neural networks
    Hu, Xinyi
    Fang, Hangxiang
    Zhang, Ling
    Zhang, Xue
    Yang, Howard H.
    Yang, Dongxiao
    Peng, Bo
    Li, Zheyang
    Hu, Haoji
    APPLIED INTELLIGENCE, 2023, 53 (16) : 19505 - 19521
  • [49] Filter Pruning via Feature Discrimination in Deep Neural Networks
    He, Zhiqiang
    Qian, Yaguan
    Wang, Yuqi
    Wang, Bin
    Guan, Xiaohui
    Gu, Zhaoquan
    Ling, Xiang
    Zeng, Shaoning
    Wang, Haijiang
    Zhou, Wujie
    COMPUTER VISION, ECCV 2022, PT XXI, 2022, 13681 : 245 - 261
  • [50] Automatic Compression Ratio Allocation for Pruning Convolutional Neural Networks
    Liu, Yunfeng
    Kong, Huihui
    Yu, Peihua
    ICVISP 2019: PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING, 2019,