DKTNet: Dual-Key Transformer Network for small object detection

被引:29
|
作者
Xu, Shoukun [1 ]
Gu, Jianan [1 ]
Hua, Yining [2 ]
Liu, Yi [1 ]
机构
[1] Changzhou Univ, Changzhou 213164, Jiangsu, Peoples R China
[2] Univ Aberdeen, Aberdeen, Scotland
基金
中国国家自然科学基金;
关键词
Small object detection; Transformer; Dual-key;
D O I
10.1016/j.neucom.2023.01.055
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Object detection is a fundamental computer vision task that plays a crucial role in a wide range of real-world applications. However, it is still a challenging task to detect the small size objects in the complex scene, due to the low resolution and noisy representation appearance caused by occlusion, distant depth view, etc. To tackle this issue, a novel transformer architecture, Dual-Key Transformer Network (DKTNet), is proposed in this paper. To improve the feature attention ability, the coherence of linear layer outputs Q and V are enhanced by a dual-K integrated from K1 and K2, which are computed along Q and V, respectively. Instead of spatial-wise attention, channel-wise self-attention mechanism is adopted to promote the important feature channels and suppress the confusing ones. Moreover, 2D and 1D convolution computations for Q, K and V are proposed. Compared with the fully-connected computa-tion in conventional transformer architectures, the 2D convolution can better capture local details and global contextual information, and the 1D convolution can reduce network complexity significantly. Experimental evaluation is conducted on both general and small object detection datasets. The superior-ity of the aforementioned features in our proposed approach is demonstrated with the comparison against the state-of-the-art approaches.(c) 2023 Elsevier B.V. All rights reserved.
引用
收藏
页码:29 / 41
页数:13
相关论文
共 50 条
  • [1] Interactive Transformer for Small Object Detection
    Wei, Jian
    Wang, Qinzhao
    Zhao, Zixu
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 77 (02): : 1699 - 1717
  • [2] Transformer-CNN for small image object detection
    Chen, Yan-Lin
    Lin, Chun-Liang
    Lin, Yu-Chen
    Chen, Tzu-Chun
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2024, 129
  • [3] TNOD: Transformer Network with Object Detection for Tag Recommendation
    Feng, Kai
    Liu, Tao
    Zhang, Heng
    Meng, Zihao
    Miao, Zemin
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 617 - 621
  • [4] Transformer guidance dual-stream network for salient object detection in optical remote sensing images
    Zhang, Yi
    Guo, Jichang
    Yue, Huihui
    Yin, Xiangjun
    Zheng, Sida
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (24) : 17733 - 17747
  • [5] Transformer guidance dual-stream network for salient object detection in optical remote sensing images
    Yi Zhang
    Jichang Guo
    Huihui Yue
    Xiangjun Yin
    Sida Zheng
    Neural Computing and Applications, 2023, 35 : 17733 - 17747
  • [6] The Optical Encoder Based on Mechanoluminescent Materials and Application in Dual-Key Encryption
    Liu, Yun
    Yu, Junwen
    Niu, Quanwang
    Wang, Xiangfu
    ADVANCED OPTICAL MATERIALS, 2025,
  • [7] Multilevel receptive field expansion network for small object detection
    Liu, Zhiwei
    Gan, Menghan
    Xiong, Li
    Mao, Xiaofeng
    Que, Yue
    IET IMAGE PROCESSING, 2023, 17 (08) : 2385 - 2398
  • [8] Consumer-Centric Insights Into Resilient Small Object Detection: SCIoU Loss and Recursive Transformer Network
    Wang, Le
    Shi, Yu
    Mao, Guojun
    Dharejo, Fayaz Ali
    Javed, Sajid
    Alathbah, Moath
    IEEE TRANSACTIONS ON CONSUMER ELECTRONICS, 2024, 70 (01) : 2178 - 2187
  • [9] Boosting Camouflaged Object Detection with Dual-Task Interactive Transformer
    Liu, Zhengyi
    Zhang, Zhili
    Tan, Yacheng
    Wu, Wei
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 140 - 146
  • [10] Feature aggregation network for small object detection
    Jing, Rudong
    Zhang, Wei
    Li, Yuzhuo
    Li, Wenlin
    Liu, Yanyan
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255