A bio-inspired positional embedding network for transformer-based models

被引:2
|
作者
Tang, Xue-song [1 ,3 ]
Hao, Kuangrong [1 ,3 ,4 ]
Wei, Hui [2 ,5 ]
机构
[1] 2999 Renmin North Rd, Shanghai 201620, Peoples R China
[2] 2005 Songhu Rd, Shanghai 200434, Peoples R China
[3] Donghua Univ, Coll Informat Sci & Technol, Shanghai, Peoples R China
[4] Minist Educ, Engn Res Ctr Digitized Text Apparel Technol, Shanghai, Peoples R China
[5] Fudan Univ, Sch Comp Sci, Lab Algorithms Cognit Models, Shanghai, Peoples R China
基金
上海市自然科学基金; 中国国家自然科学基金;
关键词
Transformers; Dorsal pathway modeling; Image classification; Position embedding; Zero padding;
D O I
10.1016/j.neunet.2023.07.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Owing to the progress of transformer-based networks, there have been significant improvements in the performance of vision models in recent years. However, there is further potential for improvement in positional embeddings that play a crucial role in distinguishing information across different positions. Based on the biological mechanisms of human visual pathways, we propose a positional embedding network that adaptively captures position information by modeling the dorsal pathway, which is responsible for spatial perception in human vision. Our proposed double-stream architecture leverages large zero-padding convolutions to learn local positional features and utilizes transformers to learn global features, effectively capturing the interaction between dorsal and ventral pathways. To evaluate the effectiveness of our method, we implemented experiments on various datasets, employing differentiated designs. Our statistical analysis demonstrates that the simple implementation significantly enhances image classification performance, and the observed trends demonstrate its biological plausibility.& COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:204 / 214
页数:11
相关论文
共 50 条
  • [1] PE-Attack: On the Universal Positional Embedding Vulnerability in Transformer-Based Models
    Gao, Shiqi
    Zhou, Haoyi
    Chen, Tianyu
    He, Mingrui
    Xu, Runhua
    Li, Jianxin
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 9359 - 9373
  • [2] Bio-Inspired Deep Spiking Neural Network for Image Classification
    Li, Jingling
    Hu, Weitai
    Yuan, Ye
    Huo, Hong
    Fang, Tao
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 294 - 304
  • [3] Adaptation of Transformer-Based Models for Depression Detection
    Adebanji, Olaronke O.
    Ojo, Olumide E.
    Calvo, Hiram
    Gelbukh, Irina
    Sidorov, Grigori
    COMPUTACION Y SISTEMAS, 2024, 28 (01): : 151 - 165
  • [4] A Transformer-Based Network for Hyperspectral Object Tracking
    Gao, Long
    Chen, Langkun
    Liu, Pan
    Jiang, Yan
    Xie, Weiying
    Li, Yunsong
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [5] Transformer-based multiple instance learning network with 2D positional encoding for histopathology image classification
    Bin Yang
    Lei Ding
    Jianqiang Li
    Yong Li
    Guangzhi Qu
    Jingyi Wang
    Qiang Wang
    Bo Liu
    Complex & Intelligent Systems, 2025, 11 (5)
  • [6] RoPIM: A Processing-in-Memory Architecture for Accelerating Rotary Positional Embedding in Transformer Models
    Jeon, Yunhyeong
    Jang, Minwoo
    Lee, Hwanjun
    Jung, Yeji
    Jung, Jin
    Lee, Jonggeon
    So, Jinin
    Kim, Daehoon
    IEEE COMPUTER ARCHITECTURE LETTERS, 2025, 24 (01) : 41 - 44
  • [7] Transformer-Based Federated Learning Models for Recommendation Systems
    Reddy, M. Sujaykumar
    Karnati, Hemanth
    Sundari, L. Mohana
    IEEE ACCESS, 2024, 12 : 109596 - 109607
  • [8] Are transformer-based models more robust than CNN-based models?
    Liu, Zhendong
    Qian, Shuwei
    Xia, Changhong
    Wang, Chongjun
    NEURAL NETWORKS, 2024, 172
  • [9] AMMU: A survey of transformer-based biomedical pretrained language models
    Kalyan, Katikapalli Subramanyam
    Rajasekharan, Ajit
    Sangeetha, Sivanesan
    JOURNAL OF BIOMEDICAL INFORMATICS, 2022, 126
  • [10] Transformer-Based Models for the Automatic Indexing of Scientific Documents in French
    Angel Gonzalez, Jose
    Buscaldi, Davide
    Sanchis, Emilio
    Hurtado, Lluis-F
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 60 - 72