A Novel Attention-Based Layer Pruning Approach for Low-Complexity Convolutional Neural Networks

被引:3
作者
Hossain, Md. Bipul [1 ]
Gong, Na [1 ]
Shaban, Mohamed [1 ]
机构
[1] Univ S Alabama, Elect & Comp Engn Dept, Mobile, AL 36688 USA
基金
美国国家科学基金会;
关键词
channel attention; deep learning; filter pruning; layer pruning;
D O I
10.1002/aisy.202400161
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning (DL) has been very successful for classifying images, detecting targets, and segmenting regions in high-resolution images such as whole slide histopathology images. However, analysis of such high-resolution images requires very high DL complexity. Several AI optimization techniques have been recently proposed that aim at reducing the complexity of deep neural networks and hence expedite their execution and eventually allow the use of low-power, low-cost computing devices with limited computation and memory resources. These methods include parameter pruning and sharing, quantization, knowledge distillation, low-rank approximation, and resource efficient architectures. Rather than pruning network structures including filters, layers, and blocks of layers based on a manual selection of a significance metric such as l1-norm and l2-norm of the filter kernels, novel highly efficient AI-driven DL optimization algorithms using variations of the squeeze and excitation in order to prune filters and layers of deep models such as VGG-16 as well as eliminate filters and blocks of residual networks such as ResNet-56 are introduced. The proposed techniques achieve significantly higher reduction in the number of learning parameters, the number of floating point operations, and memory space as compared to the-state-of-the-art methods. Herein, AI-inspired attention-based filter and layer pruning methods for extensively reducing the number of learning parameters, memory units, floating-point operations, and computational time of deep learning (DL) models as compared to the-state-of-the-art structural pruning techniques are introduced. This facilitates the realization of DL on resource-constrained edge devices and expedites the analysis of high-resolution images.image (c) 2024 WILEY-VCH GmbH
引用
收藏
页数:13
相关论文
共 50 条
[31]   Convolutional Attention-Based Bidirectional Recurrent Neural Network for Human Action Recognition [J].
Mahamkali, Aditya ;
Gali, Manvitha ;
Jena, Soumya Ranjan ;
Sreenivas, Velagapudi .
COMPUTATIONAL INTELLIGENCE, 2025, 41 (02)
[32]   A Low-Complexity Combined Encoder-LSTM-Attention Networks for EEG-based Depression Detection [J].
Ali, Noor Faris ;
Albastaki, Nabil ;
Belkacem, Abdelkader Nasreddine ;
Elfadel, Ibrahim M. ;
Atef, Mohamed .
IEEE ACCESS, 2024, 12 :129390-129403
[33]   Residential Appliance Detection Using Attention-based Deep Convolutional Neural Network [J].
Deng, Chunyu ;
Wu, Kehe ;
Wang, Binbin .
CSEE JOURNAL OF POWER AND ENERGY SYSTEMS, 2022, 8 (02) :621-633
[34]   Low-Complexity Angular Intra-Prediction Convolutional Neural Network for Lossless HEVC [J].
Huang, Hongyue ;
Schiopu, Ionut ;
Munteanu, Adrian .
2020 IEEE 22ND INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2020,
[35]   Fpar: filter pruning via attention and rank enhancement for deep convolutional neural networks acceleration [J].
Chen, Yanming ;
Wu, Gang ;
Shuai, Mingrui ;
Lou, Shubin ;
Zhang, Yiwen ;
An, Zhulin .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) :2973-2985
[36]   Attention-Based Two-Stream Convolutional Networks for Face Spoofing Detection [J].
Chen, Haonan ;
Hu, Guosheng ;
Lei, Zhen ;
Chen, Yaowu ;
Robertson, Neil M. ;
Li, Stan Z. .
IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2020, 15 :578-593
[37]   Low Complexity Convolutional Neural Networks for Wireless Receiver Chain Optimization [J].
Radi, Mohammed ;
Matus, Emil ;
Fettweis, Gerhard .
2020 IEEE INTERNATIONAL CONFERENCE ON ADVANCED NETWORKS AND TELECOMMUNICATIONS SYSTEMS (IEEE ANTS), 2020,
[38]   Gigapixel Histopathological Image Analysis Using Attention-Based Neural Networks [J].
Brancati, Nadia ;
De Pietro, Giuseppe ;
Riccio, Daniel ;
Frucci, Maria .
IEEE ACCESS, 2021, 9 :87552-87562
[39]   Classifying reservoir facies using attention-based residual neural networks [J].
Nguyen, An Hai ;
Nguyen, Khang ;
Mai, Nga .
PEERJ COMPUTER SCIENCE, 2025, 11
[40]   Interpreting sarcasm on social media using attention-based neural networks [J].
Keivanlou-Shahrestanaki, Zahra ;
Kahani, Mohsen ;
Zarrinkalam, Fattane .
KNOWLEDGE-BASED SYSTEMS, 2022, 258