Interleaved Structured Sparse Convolutional Neural Networks

被引:86
作者
Xie, Guotian [1 ,2 ]
Wang, Jingdong [3 ]
Zhang, Ting [3 ]
Lai, Jianhuang [1 ,2 ]
Hong, Richang [4 ]
Qi, Guo-Jun [5 ]
机构
[1] Sun Yat Sen Univ, Guangzhou, Guangdong, Peoples R China
[2] Guangdong Key Lab Informat Secur Technol, Guangzhou, Guangdong, Peoples R China
[3] Microsoft Res, Redmond, WA 98052 USA
[4] Hefei Univ Technol, Hefei, Anhui, Peoples R China
[5] Univ Cent Florida, Orlando, FL 32816 USA
来源
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR) | 2018年
关键词
D O I
10.1109/CVPR.2018.00922
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we study the problem of designing efficient convolutional neural network architectures with the interest in eliminating the redundancy in convolution kernels. In addition to structured sparse kernels, low-rank kernels and the product of low-rank kernels, the product of structured sparse kernels, which is a framework for interpreting the recently-developed interleaved group convolutions (IGC) and its variants (e.g., Xception), has been attracting increasing interests. Motivated by the observation that the convolutions contained in a group convolution in IGC can be further decomposed in the same manner, we present a modularized building block, IGC-V2: interleaved structured sparse convolutions. It generalizes interleaved group convolutions, which is composed of two structured sparse kernels, to the product of more structured sparse kernels, further eliminating the redundancy. We present the complementary condition and the balance condition to guide the design of structured sparse kernels, obtaining a balance among three aspects: model size, computation complexity and classification accuracy. Experimental results demonstrate the advantage on the balance among these three aspects compared to interleaved group convolutions and Xception, and competitive performance compared to other state-of-the-art architecture design methods.
引用
收藏
页码:8847 / 8856
页数:10
相关论文
共 47 条
[31]  
Lee J, 2016, J MACH LEARN RES, V17
[32]  
Li H., 2016, PRUNING FILTERS EFFI, DOI DOI 10.48550/ARXIV.1608.08710,CORR,ABS/1608.08710
[33]   Learning Efficient Convolutional Networks through Network Slimming [J].
Liu, Zhuang ;
Li, Jianguo ;
Shen, Zhiqiang ;
Huang, Gao ;
Yan, Shoumeng ;
Zhang, Changshui .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :2755-2763
[34]   ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression [J].
Luo, Jian-Hao ;
Wu, Jianxin ;
Lin, Weiyao .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5068-5076
[35]  
Mamalet Franck, 2012, Artificial Neural Networks and Machine Learning - ICANN 2012. 22nd International Conference on Artificial Neural Networks, P58, DOI 10.1007/978-3-642-33266-1_8
[36]  
Mao Huizi, 2017, ARXIV PREPRINT ARXIV
[37]  
Park Jongsoo, 2016, FASTER CNNS DIRECT S
[38]   XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks [J].
Rastegari, Mohammad ;
Ordonez, Vicente ;
Redmon, Joseph ;
Farhadi, Ali .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :525-542
[39]   ImageNet Large Scale Visual Recognition Challenge [J].
Russakovsky, Olga ;
Deng, Jia ;
Su, Hao ;
Krause, Jonathan ;
Satheesh, Sanjeev ;
Ma, Sean ;
Huang, Zhiheng ;
Karpathy, Andrej ;
Khosla, Aditya ;
Bernstein, Michael ;
Berg, Alexander C. ;
Fei-Fei, Li .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 115 (03) :211-252
[40]  
Singh Saurabh, 2016, Advances in Neural Information Processing Systems, V29