A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks

被引:2
|
作者
Hossain, Md. Bipul [1 ]
Gong, Na [1 ]
Shaban, Mohamed [1 ]
机构
[1] Univ S Alabama, Elect & Comp Engn Dept, Mobile, AL 36688 USA
基金
美国国家科学基金会;
关键词
channel attention; deep learning; filter pruning; layer pruning;
D O I
10.1002/aisy.202400161
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning (DL) has been very successful for classifying images, detecting targets, and segmenting regions in high-resolution images such as whole slide histopathology images. However, analysis of such high-resolution images requires very high DL complexity. Several AI optimization techniques have been recently proposed that aim at reducing the complexity of deep neural networks and hence expedite their execution and eventually allow the use of low-power, low-cost computing devices with limited computation and memory resources. These methods include parameter pruning and sharing, quantization, knowledge distillation, low-rank approximation, and resource efficient architectures. Rather than pruning network structures including filters, layers, and blocks of layers based on a manual selection of a significance metric such as l1-norm and l2-norm of the filter kernels, novel highly efficient AI-driven DL optimization algorithms using variations of the squeeze and excitation in order to prune filters and layers of deep models such as VGG-16 as well as eliminate filters and blocks of residual networks such as ResNet-56 are introduced. The proposed techniques achieve significantly higher reduction in the number of learning parameters, the number of floating point operations, and memory space as compared to the-state-of-the-art methods. Herein, AI-inspired attention-based filter and layer pruning methods for extensively reducing the number of learning parameters, memory units, floating-point operations, and computational time of deep learning (DL) models as compared to the-state-of-the-art structural pruning techniques are introduced. This facilitates the realization of DL on resource-constrained edge devices and expedites the analysis of high-resolution images.image (c) 2024 WILEY-VCH GmbH
引用
收藏
页数:13
相关论文
共 50 条
  • [1] A Low-Complexity Modified ThiNet Algorithm for Pruning Convolutional Neural Networks
    Tofigh, Sadegh
    Ahmad, M. Omair
    Swamy, M. N. S.
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1012 - 1016
  • [2] Low-Complexity Approximate Convolutional Neural Networks
    Cintra, Renato J.
    Duffner, Stefan
    Garcia, Christophe
    Leite, Andre
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (12) : 5981 - 5992
  • [3] Attention-based Convolutional Neural Networks for Sentence Classification
    Zhao, Zhiwei
    Wu, Youzheng
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 705 - 709
  • [4] Causal Discovery with Attention-Based Convolutional Neural Networks
    Nauta, Meike
    Bucur, Doina
    Seifert, Christin
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2019, 1 (01):
  • [5] LOW-COMPLEXITY SCALER BASED ON CONVOLUTIONAL NEURAL NETWORKS FOR ADAPTIVE VIDEO STREAMING
    Kim, Jaehwan
    Kim, Dongkyu
    Park, Min Woo
    Lee, Chaeeun
    Park, Youngo
    Choi, Kwang Pyo
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 131 - 135
  • [6] A Low-complexity Visual Tracking Approach with Single Hidden Layer Neural Networks
    Dai, Liang
    Zhu, Yuesheng
    Luo, Guibo
    He, Chao
    2014 13TH INTERNATIONAL CONFERENCE ON CONTROL AUTOMATION ROBOTICS & VISION (ICARCV), 2014, : 810 - 814
  • [7] Pre-Defined Sparsity for Low-Complexity Convolutional Neural Networks
    Kundu, Souvik
    Nazemi, Mahdi
    Pedram, Massoud
    Chugg, Keith M.
    Beerel, Peter A.
    IEEE TRANSACTIONS ON COMPUTERS, 2020, 69 (07) : 1045 - 1058
  • [8] Flattening Layer Pruning in Convolutional Neural Networks
    Jeczmionek, Ernest
    Kowalski, Piotr A.
    SYMMETRY-BASEL, 2021, 13 (07):
  • [9] Discriminative Layer Pruning for Convolutional Neural Networks
    Jordao, Artur
    Lie, Maiko
    Schwartz, William Robson
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (04) : 828 - 837
  • [10] A Low-complexity Neural BP Decoder with Network Pruning
    Han, Seokju
    Ha, Jeongseok
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 1098 - 1100