An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections

被引:165
作者
Cheng, Yu [1 ,3 ]
Yu, Felix X. [2 ,4 ]
Feris, Rogerio S. [1 ]
Kumar, Sanjiv [2 ]
Choudhary, Alok [3 ]
Chang, Shih-Fu [4 ]
机构
[1] IBM Res, Haifa, Israel
[2] Google Res, Menlo Pk, CA USA
[3] Northwestern Univ, Evanston, IL 60208 USA
[4] Columbia Univ, New York, NY 10027 USA
来源
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2015年
关键词
D O I
10.1109/ICCV.2015.327
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We explore the redundancy of parameters in deep neural networks by replacing the conventional linear projection in fully-connected layers with the circulant projection. The circulant structure substantially reduces memory footprint and enables the use of the Fast Fourier Transform to speed up the computation. Considering a fully-connected neural network layer with d input nodes, and d output nodes, this method improves the time complexity from O(d(2)) to O(d log d) and space complexity from O(d(2)) to O(d). The space savings are particularly important for modern deep convolutional neural network architectures, where fully-connected layers typically contain more than 90% of the network parameters. We further show that the gradient computation and optimization of the circulant projections can be performed very efficiently. Our experiments on three standard datasets show that the proposed approach achieves this significant gain in storage and efficiency with minimal increase in error rate compared to neural networks with unstructured projections.
引用
收藏
页码:2857 / 2865
页数:9
相关论文
共 45 条
  • [1] [Anonymous], 2013, NIPS
  • [2] [Anonymous], 2014, ECCV
  • [3] [Anonymous], MEMORY BOUNDED DEEP
  • [4] [Anonymous], 2015, CVPR
  • [5] [Anonymous], 2 INT C LEARN REPR
  • [6] [Anonymous], 2012, ECCV
  • [7] [Anonymous], 2012, ARXIV E PRINTS
  • [8] [Anonymous], 2014, ARXIV14127149
  • [9] [Anonymous], ICASSP
  • [10] [Anonymous], CIRC SYST ISCAS P 20