Visual place recognition method based on parallel omni-dimensional-dimensional dynamic attention mechanism

被引:0
作者
Liu, Peijin [1 ]
Liu, Shujie [1 ]
He, Lin [2 ]
Peng, Lijun [1 ]
Fu, Xuefeng [1 ]
机构
[1] Xian Univ Architecture & Technol, Sch Mech & Elect Engn, Xian 710055, Peoples R China
[2] Xian Univ Architecture & Technol, Fac Sci, Xian 710055, Peoples R China
关键词
visual place recognition; environmental robustness; deep learning; parallel omni-dimensional-dimensional dynamic attention; parallel strategy; MODEL;
D O I
10.37188/CJLCD.2023-0328
中图分类号
O7 [晶体学];
学科分类号
0702 ; 070205 ; 0703 ; 080501 ;
摘要
To address the issue of low robustness in visual place recognition due to environmental changes like weather'season and lighting'we propose a solution called parallel omnidimensional dynamic attention (PODAttention). ). In order to achieve dynamic and fine-grained-grained exploration of convolutional kernels across all dimensions and enhance the feature extraction network's 's ability to capture invariant features like buildings'a complementary attention mechanism is incorporated into the omni-dimensional-dimensional dynamic convolutional block. This mechanism operates on all dimensions of the convolutional kernels'including input/output channels'convolutional space and kernel quantity'enabling comprehensive attention across the entire kernel space. Furthermore'the parallel fusion of the 1x1 convolution'skip squeeze-and-excitation (SSE ) module and omni-dimensional-dimensional dynamic convolutional block yields notable benefits in terms of both feature extraction speed and the expansion of the receptive field within the visual place recognition network. By combining these components in parallel'the network gains the ability to capture more comprehensive information'resulting in enhanced accuracy for visual place recognition tasks. Experiments conducted on public datasets show that the visual place recognition method based on VGG16 and Patch-NetVLAD-NetVLAD feature aggregation improved by the POD attention mechanism'achieves 9. 7% increase in Recall@1 on the Nordland dataset and 1. 8% increase on the Mapillary Street-Level Sequences dataset. These results demonstrate that the proposed POD attention mechanism effectively enhances the robustness of visual place recognition in different environmental conditions'laying a foundation for more accurate visual localization and map construction in visual SLAM.
引用
收藏
页码:1233 / 1242
页数:10
相关论文
共 22 条
  • [1] ALI-BEY A, 2022, P 33 BRIT MACHINE VI
  • [2] GSV-CITIES: Toward appropriate supervised visual place recognition
    Ali-bey, Amar
    Chaib-draa, Brahim
    Giguere, Philippe
    [J]. NEUROCOMPUTING, 2022, 513 : 194 - 203
  • [3] Arandjelovic R, 2018, IEEE T PATTERN ANAL, V40, P1437, DOI [10.1109/CVPR.2016.572, 10.1109/TPAMI.2017.2711011]
  • [4] Visual Vocabulary with a Semantic Twist
    Arandjelovic, Relja
    Zisserman, Andrew
    [J]. COMPUTER VISION - ACCV 2014, PT I, 2015, 9003 : 178 - 195
  • [5] BAI D D, 2019, Research on deep learning based visual place recognition D.
  • [6] Bingyi Cao, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12365), P726, DOI 10.1007/978-3-030-58565-5_43
  • [7] Learning Context Flexible Attention Model for Long-Term Visual Place Recognition
    Chen, Zetao
    Liu, Lingqiao
    Sa, Inkyu
    Ge, Zongyuan
    Chli, Margarita
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 4015 - 4022
  • [8] GOYAL A, 2022, P 36 INT C NEURAL IN
  • [9] Patch-NetVLAD: Multi-Scale Fusion of Locally-Global Descriptors for Place Recognition
    Hausler, Stephen
    Garg, Sourav
    Xu, Ming
    Milford, Michael
    Fischer, Tobias
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14136 - 14147
  • [10] Hou Y, 2015, 2015 IEEE INTERNATIONAL CONFERENCE ON INFORMATION AND AUTOMATION, P2238, DOI 10.1109/ICInfA.2015.7279659