An improved interaction-and-aggregation network for person re-identification

被引:1
作者
Tao, Huanjie [1 ,2 ,3 ]
Bao, Wenjie [1 ]
Duan, Qianyue [1 ]
Hu, Zhenwu [1 ]
An, Jianfeng [1 ,2 ,3 ]
Xie, Chao [4 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Xian 710129, Peoples R China
[2] Northwestern Polytech Univ, Engn & Res Ctr Embedded Syst Integrat, Minist Educ, Xian 710129, Peoples R China
[3] Natl Engn Lab Integrated Aerosp Ground Ocean Big D, Xian 710129, Peoples R China
[4] Nanjing Forestry Univ, Coll Mech & Elect Engn, Nanjing 210037, Peoples R China
基金
中国国家自然科学基金;
关键词
Person Re-identification; Gating Mechanism; Context Information; Attention Mechanism;
D O I
10.1007/s11042-023-15531-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Person re-identification (ReID) aims to match a specific person across non-overlapping camera views and has wide application prospects. However, existing methods are still susceptible to occlusion and missing critical parts. Most methods fuse low-level detail features and high-level strong semantic features using feature concatenation or addition, leading to useful information being overwhelmed by a large amount of useless information. In addition, many methods extract spatial context features by designing different blocks but ignore the local channel context features. To relieve these issues, this paper presents an improved interaction-and-aggregation network (IIANet) to learn more representative feature representation. First, to improve model robustness to serious occlusion or missing crucial parts of the target person, we employ a global multi-scale module (MSM) to extract multi-scale features by multi-branch convolution and hierarchical residual connection. Second, to selectively fuse low-level detail features and high-level semantic features effectively, we design a gated fully fusion module (GFFM) to control information transmission and reduce feature interferences in fusing different-level features. Finally, we adopt a channel context module (CCM) to learn channel context information via multi-scale local fusion. Sufficient experiments demonstrate the better performances of our IIANet on dataset Market-1501. The mAP and Rank-1 accuracy of our model reach 84.9% and 94.2%, respectively. Our code is available at: https://gitee.com/bingsfan/iianet/tree/master/
引用
收藏
页码:44053 / 44069
页数:17
相关论文
共 54 条
  • [1] Toward image super-resolution based on local regression and nonlocal means
    Bastanfard, Azam
    Amirkhani, Dariush
    MohammadinAff, Mohammad
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (16) : 23473 - 23492
  • [2] Deep Reinforcement Learning for Channel and Power Allocation in UAV-enabled IoT Systems
    Cao, Yang
    Zhang, Lin
    Liang, Ying-Chang
    [J]. 2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [3] Cho K, 2020, ARXIV
  • [4] Fan Xing, 2018, AS C COMP VIS, P19, DOI DOI 10.1007/978-3-030-20890-5_2
  • [5] Gao Shang, 2020, CVPR, P11744
  • [6] Res2Net: A New Multi-Scale Backbone Architecture
    Gao, Shang-Hua
    Cheng, Ming-Ming
    Zhao, Kai
    Zhang, Xin-Yu
    Yang, Ming-Hsuan
    Torr, Philip
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (02) : 652 - 662
  • [7] Deep Metric Learning with Hierarchical Triplet Loss
    Ge, Weifeng
    Huang, Weilin
    Dong, Dengke
    Scott, Matthew R.
    [J]. COMPUTER VISION - ECCV 2018, PT VI, 2018, 11210 : 272 - 288
  • [8] Ge Y, 2018, IEEE C EVOL COMPUTAT, V31
  • [9] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [10] Foreground-aware Pyramid Reconstruction for Alignment-free Occluded Person Re-identification
    He, Lingxiao
    Wang, Yinggang
    Liu, Wu
    Zhao, He
    Sun, Zhenan
    Feng, Jiashi
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 8449 - 8458