Extract Generalization Ability from Convolutional Neural Networks

被引:0
作者
Wu, Huan [1 ]
Wu, JunMin [1 ]
Ding, Jie [1 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei, Peoples R China
来源
2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2018年
关键词
Convolutional neural networks; redundancy; compression; acceleration; DEEP NETWORK;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the development of deep learning, the trend of mobile artificial intelligent is emerging at a rapid pace. Over time, a series of mobile devices such as smart speakers and smartphones gradually have become the first choice of artificial intelligent landing. However, the hardware resources of mobile devices are very limited. In terms of time, memory, or energy consumption, large neural networks cannot be deployed on mobile devices. Therefore, it is necessary to explore a new compression method. Studies have shown that much redundancy exists in convolutional neural networks. This means that the neural network structure can be trimmed without affecting the accuracy. Inspired by the autoencoder, this paper presents a new compression method. The method can remove redundant neurons and convolution kernels. It extracts the generalization ability of convolutional neural networks onto smaller models. Experimental results show that the pruning rate is approximately 4x to 21x, and the speedup is approximately 2x to 5x.
引用
收藏
页码:729 / 734
页数:6
相关论文
共 31 条
  • [1] [Anonymous], 2016, ARXIV160202830
  • [2] [Anonymous], 2014, Learning Semantic Image Representations at a Large Scale
  • [3] Baldi P., 2012, P ICML WORKSH UNS TR, V27, P37, DOI DOI 10.1561/2200000006
  • [4] Xception: Deep Learning with Depthwise Separable Convolutions
    Chollet, Francois
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1800 - 1807
  • [5] Courbariaux M., 2015, ADV NEURAL INFORM PR, P3105
  • [6] Dally WJ, 2015, ARXIV151000149, V2
  • [7] Denil M., 2013, P ADV NEUR INF PROC, P2148, DOI DOI 10.5555/2999792.2999852
  • [8] Denton EL, 2014, Advances in neural information processing systems, P1269
  • [9] Graham B., 2014, Fractional Max-Pooling
  • [10] Gupta S, 2015, PR MACH LEARN RES, V37, P1737