Adder Attention for Vision Transformer

被引:0
作者
Shu, Han [1 ]
Wang, Jiahao [2 ]
Chen, Hanting [1 ,3 ]
Li, Lin [4 ]
Yang, Yujiu [2 ]
Wang, Yunhe [1 ]
机构
[1] Huawei Noahs Ark Lab, Hong Kong, Peoples R China
[2] Tsinghua Shenzhen Int Grad Sch, Shenzhen, Peoples R China
[3] Peking Univ, Beijing, Peoples R China
[4] Huawei Technol, Shenzhen, Peoples R China
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021) | 2021年 / 34卷
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformer is a new kind of calculation paradigm for deep learning which has shown strong performance on a large variety of computer vision tasks. However, compared with conventional deep models (e.g., convolutional neural networks), vision transformers require more computational resources which cannot be easily deployed on mobile devices. To this end, we present to reduce the energy consumptions using adder neural network (AdderNet). We first theoretically analyze the mechanism of self-attention and the difficulty for applying adder operation into this module. Specifically, the feature diversity, i.e., the rank of attention map using only additions cannot be well preserved. Thus, we develop an adder attention layer that includes an additional identity mapping. With the new operation, vision transformers constructed using additions can also provide powerful feature representations. Experimental results on several benchmarks demonstrate that the proposed approach can achieve highly competitive performance to that of the baselines while achieving an about 2(similar to)3x reduction on the energy consumption.
引用
收藏
页数:11
相关论文
共 50 条
[41]   ASAFormer: Visual tracking with convolutional vision transformer and asymmetric selective attention [J].
Gong, Xiaomei ;
Zhang, Yi ;
Hu, Shu .
Knowledge-Based Systems, 2024, 291
[42]   Lightweight Vision Transformer with Spatial and Channel Enhanced Self-Attention [J].
Zheng, Jiahao ;
Yang, Longqi ;
Li, Yiying ;
Yang, Ke ;
Wang, Zhiyuan ;
Zhou, Jun .
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, :1484-1488
[43]   DHFormer: A Vision Transformer-Based Attention Module for Image Dehazing [J].
Wasi, Abdul ;
Shiney, O. Jeba .
COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT I, 2024, 2009 :148-159
[44]   POLARIMETRIC HRRP RECOGNITION USING VISION TRANSFORMER WITH POLARIMETRIC PREPROCESSING AND ATTENTION LOSS [J].
Gao, Fan ;
Ren, Dawei ;
Yin, Junjun ;
Yang, Jian .
2024 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2024), 2024, :10838-10842
[45]   A Vision Transformer with Adaptive Cross-Image and Cross-Resolution Attention [J].
Murray, Benjamin A. K. ;
Tan, Wei R. ;
Canas, Liane S. ;
Smith, Catherine H. ;
Mahil, Satveer K. ;
Ourselin, Sebastien ;
Modat, Marc .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION-MICCAI 2024 WORKSHOPS, ISIC 2024, IMIMIC 2024, EARTH 2024, DECAF 2024, 2025, 15274 :35-44
[46]   PlantViQ: Disease Recognition Across Varied Environments with Vision Transformer and Quadrangle Attention [J].
Li, Shuting ;
Chen, Baoyu ;
Li, Feng ;
He, Jingmei ;
He, Feiyong ;
Hu, Yingbiao ;
Chen, Jingjia ;
Li, Huinian .
ADVANCED INTELLIGENT COMPUTING IN BIOINFORMATICS, PT I, ICIC 2024, 2024, 14881 :441-452
[47]   A novel hybrid attention gate based on vision transformer for the detection of surface defects [J].
Uzen, Hueseyin ;
Turkoglu, Muammer ;
Ozturk, Dursun ;
Hanbay, Davut .
SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (10) :6835-6851
[48]   A novel twin vision transformer framework for crop disease classification with deformable attention [J].
Padshetty, Smitha ;
Ambika .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 105
[49]   EViT: An Eagle Vision Transformer With Bi-Fovea Self-Attention [J].
Shi, Yulong ;
Sun, Mingwei ;
Wang, Yongshuai ;
Ma, Jiahao ;
Chen, Zengqiang .
IEEE TRANSACTIONS ON CYBERNETICS, 2025, 55 (03) :1288-1300
[50]   ReViT: Vision Transformer Accelerator With Reconfigurable Semantic-Aware Differential Attention [J].
Zou, Xiaofeng ;
Chen, Cen ;
Shao, Hongen ;
Wang, Qinyu ;
Zhuang, Xiaobin ;
Li, Yangfan ;
Li, Keqin .
IEEE TRANSACTIONS ON COMPUTERS, 2025, 74 (03) :1079-1093