A Lightweight Block With Information Flow Enhancement for Convolutional Neural Networks

被引:7
作者
Bao, Zhiqiang [1 ]
Yang, Shunzhi [1 ]
Huang, Zhenhua [1 ]
Zhou, MengChu [2 ,3 ]
Chen, Yunwen [4 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510631, Peoples R China
[2] New Jersey Inst Technol, Dept Elect & Comp Engn, Newark, NJ 07102 USA
[3] St Petersburg State Marine Tech Univ, Dept Cyber Phys Syst, St Petersburg 198262, Russia
[4] DataGrand Inc, Res & Dev Dept, Shanghai 201203, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolutional neural network; lightweight; information flow; activation function; affine transformation;
D O I
10.1109/TCSVT.2023.3237615
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional neural networks (CNNs) have demonstrated excellent capability in various visual recognition tasks but impose an excessive computational burden. The latter problem is commonly solved by utilizing lightweight sparse networks. However, such networks have a limited receptive field in a few layers, and the majority of these networks face a severe information barrage due to their sparse structures. Spurred by these deficiencies, this work proposes a Squeeze Convolution block with Information Flow Enhancement (SCIFE), comprising a Divide-and-Squeeze Convolution and an Information Flow Enhancement scheme. The former module constructs a multi-layer structure through multiple squeeze operations to increase the receptive field and reduce computation. The latter replaces the affine transformation with the point convolution and dynamically adjusts the activation function's threshold, enhancing information flow in both channels and layers. Moreover, we reveal that the original affine transformation may harm the network's generalization capability. To overcome this issue, we utilize a point convolution with a zero initial mean. SCIFE can serve as a plug-and-play replacement for vanilla convolution blocks in mainstream CNNs, while extensive experimental results demonstrate that CNNs equipped with SCIFE compress benchmark structures without sacrificing performance, outperforming their competitors.
引用
收藏
页码:3570 / 3584
页数:15
相关论文
共 75 条
  • [1] Barron J. T., 2017, PROC INT C LEARN REP, P763
  • [2] Clevert D.-A., 2015, ARXIV
  • [3] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [4] Weighted Feature Fusion of Convolutional Neural Network and Graph Attention Network for Hyperspectral Image Classification
    Dong, Yanni
    Liu, Quanwei
    Du, Bo
    Zhang, Liangpei
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1559 - 1572
  • [5] Dugas C, 2001, ADV NEUR IN, V13, P472
  • [6] Eger S, 2013, J INTEGER SEQ, V16
  • [7] The Pascal Visual Object Classes (VOC) Challenge
    Everingham, Mark
    Van Gool, Luc
    Williams, Christopher K. I.
    Winn, John
    Zisserman, Andrew
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) : 303 - 338
  • [8] Howard AG, 2017, Arxiv, DOI [arXiv:1704.04861, DOI 10.48550/ARXIV.1704.04861, 10.48550/arXiv.1704.04861]
  • [9] Glorot X., 2011, PROC 14 INT C ARTIF, V15, P315
  • [10] Gulcehre C, 2016, PR MACH LEARN RES, V48