A Lightweight Transformer with Convolutional Attention

被引:0
作者
Zeng, Kungan [1 ]
Paik, Incheon [1 ]
机构
[1] Univ Aizu, Sch Comp Sci & Engn, Fukushima, Japan
来源
2020 11TH INTERNATIONAL CONFERENCE ON AWARENESS SCIENCE AND TECHNOLOGY (ICAST) | 2020年
关键词
neural machine translation; Transformer; CNN; Muti-head attention;
D O I
10.1109/ICAST51195.2020.9319489
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Neural machine translation (NMT) goes through rapid development because of the application of various deep learning techs. Especially, how to construct a more effective structure of NMT attracts more and more attention. Transformer is a state-of-the-art architecture in NMT. It replies on the self-attention mechanism exactly instead of recurrent neural networks (RNN). The Multi-head attention is a crucial part that implements the self-attention mechanism, and it also dramatically affects the scale of the model. In this paper, we present a new Multi-head attention by combining convolution operation. In comparison with the base Transformer, our approach can reduce the number of parameters effectively. And we perform a reasoned experiment. The result shows that the performance of the new model is similar to the base model.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Improved SwinUNet with fusion transformer and large kernel convolutional attention for liver and tumor segmentation in CT images
    Linfeng Jiang
    Jiani Hu
    Tongyuan Huang
    Scientific Reports, 15 (1)
  • [32] RT-CBAM: Refined Transformer Combined with Convolutional Block Attention Module for Underwater Image Restoration
    Ye, Renchuan
    Qian, Yuqiang
    Huang, Xinming
    SENSORS, 2024, 24 (18)
  • [33] MCANet: Hierarchical cross-fusion lightweight transformer based on multi-ConvHead attention for object detection
    Zhao, Zuopeng
    Hao, Kai
    Liu, Xiaofeng
    Zheng, Tianci
    Xu, Junjie
    Cui, Shuya
    He, Chen
    Zhou, Jie
    Zhao, Guangming
    IMAGE AND VISION COMPUTING, 2023, 136
  • [34] SSformer: A Lightweight Transformer for Semantic Segmentation
    Shi, Wentao
    Xu, Jing
    Gao, Pan
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [35] Bearing fault diagnosis using transfer learning and self-attention ensemble lightweight convolutional neural network
    Zhong, Hongyu
    Lv, Yong
    Yuan, Rui
    Yang, Di
    NEUROCOMPUTING, 2022, 501 : 765 - 777
  • [36] Global texture sensitive convolutional transformer for medical image steganalysis
    Zhou, Zhengyuan
    Chen, Kai
    Hu, Dianlin
    Shu, Huazhong
    Coatrieux, Gouenou
    Coatrieux, Jean Louis
    Chen, Yang
    MULTIMEDIA SYSTEMS, 2024, 30 (03)
  • [37] Dual attention transformer network for hyperspectral image classification
    Shu, Zhenqiu
    Wang, Yuyang
    Yu, Zhengtao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
  • [38] Convolutional transformer network for fine-grained action recognition
    Ma, Yujun
    Wang, Ruili
    Zong, Ming
    Ji, Wanting
    Wang, Yi
    Ye, Baoliu
    NEUROCOMPUTING, 2024, 569
  • [39] Convolutional transformer network for paranasal anomaly classification in the maxillary sinus
    Bhattacharya, Debayan
    Behrendt, Finn
    Maack, Lennart
    Becker, Benjamin Tobias
    Beyersdorff, Dirk
    Petersen, Elina
    Petersen, Marvin
    Cheng, Bastian
    Eggert, Dennis
    Betz, Christian
    Hoffmann, Anna Sophie
    Schlaefer, Alexander
    COMPUTER-AIDED DIAGNOSIS, MEDICAL IMAGING 2024, 2024, 12927
  • [40] Attention-guided hybrid transformer-convolutional neural network for underwater image super-resolution
    Zhan, Zihan
    Li, Chaofeng
    Zhang, Yuqi
    JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (01)