Enhanced mechanisms of pooling and channel attention for deep learning feature maps

被引:2
|
作者
Li, Hengyi [1 ]
Yue, Xuebin [1 ]
Meng, Lin [2 ]
机构
[1] Ritsumeikan Univ, Grad Sch Sci & Engn, Kusatsu, Shiga, Japan
[2] Ritsumeikan Univ, Coll Sci & Engn, Kusatsu, Shiga, Japan
关键词
DNNs; Max pooling; Average pooling; FMAPooling; Self-attention; FMAttn;
D O I
10.7717/peerj-cs.1161
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The pooling function is vital for deep neural networks (DNNs). The operation is to generalize the representation of feature maps and progressively cut down the spatial size of feature maps to optimize the computing consumption of the network. Furthermore, the function is also the basis for the computer vision attention mechanism. However, as a matter of fact, pooling is a down-sampling operation, which makes the feature-map representation approximately to small translations with the summary statistic of adjacent pixels. As a result, the function inevitably leads to information loss more or less. In this article, we propose a fused max-average pooling (FMAPooling) operation as well as an improved channel attention mechanism (FMAttn) by utilizing the two pooling functions to enhance the feature representation for DNNs. Basically, the methods are to enhance multiple-level features extracted by max pooling and average pooling respectively. The effectiveness of the proposals is verified with VGG, ResNet, and MobileNetV2 architectures on CIFAR10/100 and ImageNet100. According to the experimental results, the FMAPooling brings up to 1.63% accuracy improvement compared with the baseline model; the FMAttn achieves up to 2.21% accuracy improvement compared with the previous channel attention mechanism. Furthermore, the proposals are extensible and could be embedded into various DNN models easily, or take the place of certain structures of DNNs. The computation burden introduced by the proposals is negligible.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Enhanced mechanisms of pooling and channel attention for deep learning feature maps
    Li H.
    Yue X.
    Meng L.
    PeerJ Computer Science, 2022, 8
  • [2] Generalized Local Attention Pooling for Deep Metric Learning
    Roig, Carlos
    Varas, David
    Masuda, Issey
    Riveiro, Juan Carlos
    Bou-Balust, Elisenda
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 9951 - 9958
  • [3] Detection of Attention Deficit Hyperactivity Disorder based on EEG feature maps and deep learning
    Cura, Ozlem Karabiber
    Akan, Aydin
    Atli, Sibel Kocaaslan
    BIOCYBERNETICS AND BIOMEDICAL ENGINEERING, 2024, 44 (03) : 450 - 460
  • [4] Fast Deep Learning Classification With Multilinear Feature Maps
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [5] Partial channel pooling attention beats convolutional attention
    Zhang, Jun
    Slamu, Wushour
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 237
  • [6] Harnessing edge-enhanced attention mechanisms for supernova detection in deep learning frameworks
    Yin, K.
    Jia, J.
    Li, F.
    Gao, X.
    Sun, T.
    ASTRONOMY AND COMPUTING, 2024, 46
  • [7] A General Survey on Attention Mechanisms in Deep Learning
    Brauwers, Gianni
    Frasincar, Flavius
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (04) : 3279 - 3298
  • [8] Neighbour feature attention-based pooling
    Li, Xiaosong
    Wu, Yanxia
    Fu, Yan
    Tang, Chuheng
    Zhang, Lidan
    NEUROCOMPUTING, 2022, 501 : 285 - 293
  • [9] DEEP COLOR IMAGE DEMOSAICKING WITH FEATURE PYRAMID CHANNEL ATTENTION
    Kang, Qi
    Fu, Ying
    Huang, Hua
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 246 - 251
  • [10] Channel Attention Enhanced Deep Network for Segmenting Exudate
    Maiti, Souvik
    Maji, Debasis
    Dhara, Ashis Kumar
    Sarkar, Gautam
    2022 IEEE 6TH INTERNATIONAL CONFERENCE ON CONDITION ASSESSMENT TECHNIQUES IN ELECTRICAL SYSTEMS, CATCON, 2022, : 94 - 98