Structured pruning of recurrent neural networks through neuron selection

被引:30
|
作者
Wen, Liangjian [1 ]
Zhang, Xuanyang [1 ]
Bai, Haoli [2 ]
Xu, Zenglin [1 ,3 ]
机构
[1] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, SMILE Lab, Chengdu 610031, Peoples R China
[2] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Shatin, Hong Kong 999077, Peoples R China
[3] Ctr Artificial Intelligence, Peng Cheng Lab, Shenzhen, Guangdong, Peoples R China
关键词
Feature selection; Recurrent neural networks; Learning sparse models; Model compression;
D O I
10.1016/j.neunet.2019.11.018
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) have recently achieved remarkable successes in a number of applications. However, the huge sizes and computational burden of these models make it difficult for their deployment on edge devices. A practically effective approach is to reduce the overall storage and computation costs of RNNs by network pruning techniques. Despite their successful applications, those pruning methods based on Lasso either produce irregular sparse patterns in weight matrices, which is not helpful in practical speedup. To address these issues, we propose a structured pruning method through neuron selection which can remove the independent neuron of RNNs. More specifically, we introduce two sets of binary random variables, which can be interpreted as gates or switches to the input neurons and the hidden neurons, respectively. We demonstrate that the corresponding optimization problem can be addressed by minimizing the L-0 norm of the weight matrix. Finally, experimental results on language modeling and machine reading comprehension tasks have indicated the advantages of the proposed method in comparison with state-of-the-art pruning competitors. In particular, nearly 20x practical speedup during inference was achieved without losing performance for the language model on the Penn TreeBank dataset, indicating the promising performance of the proposed method. (C) 2019 Elsevier Ltd. All rights reserved.
引用
收藏
页码:134 / 141
页数:8
相关论文
共 50 条
  • [21] Recognizing recurrent neural networks (rRNN): Bayesian inference for recurrent neural networks
    Bitzer, Sebastian
    Kiebel, Stefan J.
    BIOLOGICAL CYBERNETICS, 2012, 106 (4-5) : 201 - 217
  • [22] Connecting weighted automata, tensor networks and recurrent neural networks through spectral learning
    Li, Tianyu
    Precup, Doina
    Rabusseau, Guillaume
    MACHINE LEARNING, 2024, 113 (05) : 2619 - 2653
  • [23] Connecting weighted automata, tensor networks and recurrent neural networks through spectral learning
    Tianyu Li
    Doina Precup
    Guillaume Rabusseau
    Machine Learning, 2024, 113 : 2619 - 2653
  • [24] Enabling Combined Relay Selection in Stochastic Wireless Networks by Recurrent Neural Computing
    Tang, Jiashen
    Dang, Shuping
    Abdullah, Salwani
    Nazri, Mohd Zakree Ahmad
    Sabar, Nasser R.
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2023, 22 (11) : 6410 - 6426
  • [25] Overview of recurrent neural networks
    Liu J.-W.
    Song Z.-Y.
    Kongzhi yu Juece/Control and Decision, 2022, 37 (11): : 2753 - 2768
  • [26] RECURRENT NEURAL NETWORKS FOR SYLLABICATION
    HUNT, A
    SPEECH COMMUNICATION, 1993, 13 (3-4) : 323 - 332
  • [27] Stability of Recurrent Neural Networks
    Jalab, Hamid A.
    Ibrahim, Rabha W.
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2006, 6 (12): : 159 - 164
  • [28] Recurrent Neural Networks for Storytelling
    Choi, YunSeok
    Kim, SuAh
    Lee, Jee-Hyong
    2016 JOINT 8TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING AND INTELLIGENT SYSTEMS (SCIS) AND 17TH INTERNATIONAL SYMPOSIUM ON ADVANCED INTELLIGENT SYSTEMS (ISIS), 2016, : 841 - 845
  • [29] PDD: Pruning Neural Networks During Knowledge Distillation
    Dan, Xi
    Yang, Wenjie
    Zhang, Fuyan
    Zhou, Yihang
    Yu, Zhuojun
    Qiu, Zhen
    Zhao, Boyuan
    Dong, Zeyu
    Huang, Libo
    Yang, Chuanguang
    COGNITIVE COMPUTATION, 2024, 16 (06) : 3457 - 3467
  • [30] Bidirectional recurrent neural networks
    Schuster, M
    Paliwal, KK
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1997, 45 (11) : 2673 - 2681