Dynamic Center Aggregation Loss With Mixed Modality for Visible-Infrared Person Re-Identification

被引:21
作者
Kong, Jun [1 ]
He, Qibin [2 ]
Jiang, Min [2 ]
Liu, Tianshan [3 ]
机构
[1] Jiangnan Univ, Minist Educ, Key Lab Adv Proc Control Light Ind, Wuxi 214122, Jiangsu, Peoples R China
[2] Jiangnan Univ, Jiangsu Prov Engn Lab Pattern Recognit & Computat, Wuxi 214122, Jiangsu, Peoples R China
[3] Hong Kong Polytech Univ, Dept Elect & Informat Engn, Hong Kong 999077, Peoples R China
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Feature extraction; Training; Mixers; Image color analysis; Bridges; Testing; Task analysis; Person re-identification; cross-modality; mixed modality; NETWORK;
D O I
10.1109/LSP.2021.3115040
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Visible-infrared person re-identification (VI-ReID) is a challenging cross-modality pedestrian retrieval task which aims to match person images between the visible and infrared modality of the same identity. Existing methods usually adopt two-stream network to solve cross-modality gap, but they ignore the pixel-level discrepancy between the visible and infrared images. Some methods introduce auxiliary modalities in the network, but they lack powerful constraints on the feature distribution of multiple modalities. In this letter, we propose a Dynamic Center Aggregation (DCA) loss with mixed modality for VI-ReID. Concretely, we employ a mixed modality as a bridge between the visible and infrared modality, reducing the difference of the two modalities at the pixel-level. The mixed modality is generated by a Dual-modality Feature Mixer (DFM), which combines the features of visible and infrared images. Moreover, we dynamically adjust the relative distance across multi-modality through DCA loss, which is conducive to explore the modality-invariant feature. We evaluate the proposed method on two public available VI-ReID datasets (SYSU-MM01 and RegDB). Experimental results demonstrate that our method achieves competitive performance.
引用
收藏
页码:2003 / 2007
页数:5
相关论文
共 34 条
  • [1] Choi S., 2020, P IEEE C COMP VIS PA, P10257
  • [2] Dai PY, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P677
  • [3] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [4] Goodfellow I., 2020, ADV NEUR IN, V63, P139, DOI [DOI 10.1145/3422622, 10.1145/3422622]
  • [5] Dual-alignment Feature Embedding for Cross-modality Person Re-identification
    Hao, Yi
    Wang, Nannan
    Gao, Xinbo
    Li, Jie
    Wang, Xiaoyu
    [J]. PROCEEDINGS OF THE 27TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA (MM'19), 2019, : 57 - 65
  • [6] Hao Y, 2019, AAAI CONF ARTIF INTE, P8385
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
  • [9] Generated Data With Sparse Regularized Multi-Pseudo Label for Person Re-Identification
    Huang, Liqin
    Yang, Qingqing
    Wu, Junyi
    Huang, Yan
    Wu, Qiang
    Xu, Jingsong
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2020, 27 (27) : 391 - 395
  • [10] Ioffe S, 2015, PR MACH LEARN RES, V37, P448