Multi-granularity feature utilization network for cross-modality visible-infrared person re-identification

被引:1
作者
Zhang, Guoqing [1 ,2 ]
Zhang, Yinyin [1 ]
Chen, Yuhao [1 ]
Zhang, Hongwei [1 ]
Zheng, Yuhui [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Comp Sci, Nanjing 210044, Peoples R China
[2] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Minist Educ, Nanjing 210044, Peoples R China
基金
中国国家自然科学基金;
关键词
Cross-modality; Person re-identification; Multi-granularity; Heterogeneous center loss;
D O I
10.1007/s00500-023-08321-7
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cross-modality visible-infrared person re-identification (VI-ReID) aims to recognize images with the same identity between visible modality and infrared modality, which is a very challenging task because it not only includes the troubles of variations between cross-cameras in traditional person ReID, but also suffers from the huge differences between two modalities. Some existing VI-ReID methods extract the modality-shared information from global features through single-stream or double-stream networks, while ignoring the complementarity between fine-grained and coarse-grained information. To solve this problem, our paper designs a multi-granularity feature utilization network (MFUN) to make up for the lack of shared features between different modalities by promoting the complementarity between coarse-grained features and fine-grained features. Firstly, in order to learn fine-grained shared features better, we design a local feature constraint module, which uses both hard-mining triplet loss and heterogeneous center loss to constrain local features in the common subspace, so as to better promote intra-class compactness and inter-class differences at the level of sample and class center. Then, our method uses a multi-modality feature aggregationmodule for global features to fuse the information of two modalities to narrowthe modality gap. Through the combination of these two modules, visible and infrared image features can be better fused, thus alleviating the problem of modality discrepancy and supplementing the lack of modality-shared information. Extensive experimental results on RegDB and SYSU-MM01 datasets fully prove that our proposed MFUN outperforms the state-of-the-art solutions. Our code is available at https://github.com/ZhangYinyinzzz/MFUN.
引用
收藏
页数:14
相关论文
共 51 条
  • [1] Neural Feature Search for RGB-Infrared Person Re-Identification
    Chen, Yehansen
    Wan, Lin
    Li, Zhihang
    Jing, Qianyan
    Sun, Zongyuan
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 587 - 597
  • [2] Choi S, 2020, PROC CVPR IEEE, P10254, DOI 10.1109/CVPR42600.2020.01027
  • [3] Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
  • [4] Deep feature learning with relative distance comparison for person re-identification
    Ding, Shengyong
    Lin, Liang
    Wang, Guangrun
    Chao, Hongyang
    [J]. PATTERN RECOGNITION, 2015, 48 (10) : 2993 - 3003
  • [5] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [6] Dual-alignment Feature Embedding for Cross-modality Person Re-identification
    Hao, Yi
    Wang, Nannan
    Gao, Xinbo
    Li, Jie
    Wang, Xiaoyu
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 57 - 65
  • [7] Hu B., 2021, PROC IEEE INT C MULT, DOI DOI 10.1109/ICMES1207.2021.9428376
  • [8] Kulis B, 2011, PROC CVPR IEEE, P1785, DOI 10.1109/CVPR.2011.5995702
  • [9] Kumar D, 2020, IEEE WINT CONF APPL, P2634, DOI 10.1109/WACV45572.2020.9093606
  • [10] A Survey of Open-World Person Re-Identification
    Leng, Qingming
    Ye, Mang
    Tian, Qi
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (04) : 1092 - 1108