Global and Part Feature Fusion for Cross-Modality Person Re-Identification

被引:1
作者
Wang, Xianju [1 ,2 ]
Cordova, Ronald S. [1 ,3 ]
机构
[1] Angeles Univ Fdn, Grad Sch, Angeles 2009, Philippines
[2] Fuyang Normal Univ, Sch Phys & Elect Engn, Fuyang 236037, Anhui, Peoples R China
[3] Gulf Coll, Fac Comp Sci, Muscat, Oman
来源
IEEE ACCESS | 2022年 / 10卷
关键词
Visible-infrared; Re-ID; feature fusion; cross-modality;
D O I
10.1109/ACCESS.2022.3222267
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Visible-Infrared person re-identification (VI Re-ID) is a challenging but practical task that aims at matching pedestrian images between the visible(daytime) modality and the infrared(nighttime) modality, playing an important role in criminal investigation and intelligent video surveillance applications. Numerous previous studies focused on alleviating the modality discrepancy and obtaining discriminating features by devising complex networks for VI Re-ID, but a cumbersome network structure is not suitable for practical industrial applications. In this paper, we propose a novel fusion method of global and part features to extract distinguishing features and alleviate cross-modality differences, named Global and Part Feature Fusion network(GPFF), which has not been well studied in the current literature. Specifically, we first adopt a dual-stream ResNet50 as a backbone network to alleviate the modality discrepancy. Then, We explore how to fuse global and local features to obtain discriminative features. Finally, we apply a heterogeneous center triplet loss(hetero-center triplet loss) instead of traditional triplet loss to guide sample center learning. Our proposed approach is simple but effective, and can remarkably boost the performance of VI Re-ID. The results of experiments on two public datasets(SYSU-MM01 and RegDB) demonstrate that our approach is superior to the state-of-the-art methods. Through experiments, we find that the effective fusion of global features and local features plays an important role in extracting discriminative features.
引用
收藏
页码:122038 / 122046
页数:9
相关论文
共 41 条
  • [1] Chen YC, 2015, PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), P3402
  • [2] Choi S, 2020, PROC CVPR IEEE, P10254, DOI 10.1109/CVPR42600.2020.01027
  • [3] Fan X, 2020, Arxiv, DOI arXiv:2003.00213
  • [4] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [5] Leaning compact and representative features for cross-modality person re-identification
    Gao, Guangwei
    Shao, Hao
    Wu, Fei
    Yang, Meng
    Yu, Yi
    [J]. WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2022, 25 (04): : 1649 - 1666
  • [6] Hao Y, 2019, AAAI CONF ARTIF INTE, P8385
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Hermans A, 2017, Arxiv, DOI arXiv:1703.07737
  • [9] ThermalGAN: Multimodal Color-to-Thermal Image Translation for Person Re-identification in Multispectral Dataset
    Kniaz, Vladimir V.
    Knyaz, Vladimir A.
    Hladuvka, Jiri
    Kropatsch, Walter G.
    Mizginov, Vladimir
    [J]. COMPUTER VISION - ECCV 2018 WORKSHOPS, PT VI, 2019, 11134 : 606 - 624
  • [10] Liao SC, 2015, PROC CVPR IEEE, P2197, DOI 10.1109/CVPR.2015.7298832