Towards a Unified Middle Modality Learning for Visible-Infrared Person Re-Identification

被引:107
|
作者
Zhang, Yukang [1 ]
Yan, Yan [1 ]
Lu, Yang [1 ]
Wang, Hanzi [1 ]
机构
[1] Xiamen Univ, Fujian Key Lab Sensing & Comp Smart City, Xiamen, Peoples R China
来源
PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021 | 2021年
基金
中国国家自然科学基金;
关键词
VI-ReID; Non-Linear; Middle Modality; Distribution Consistency;
D O I
10.1145/3474085.3475250
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visible-infrared person re-identification (VI-ReID) aims to search identities of pedestrians across different spectra. In this task, one of the major challenges is the modality discrepancy between the visible (VIS) and infrared (IR) images. Some state-of-the-art methods try to design complex networks or generative methods to mitigate the modality discrepancy while ignoring the highly non-linear relationship between the two modalities of VIS and IR. In this paper, we propose a non-linear middle modality generator (MMG), which helps to reduce the modality discrepancy. Our MMG can effectively project VIS and IR images into a unified middle modality image (UMMI) space to generate middle-modality (M-modality) images. The generated M-modality images and the original images are fed into the backbone network to reduce the modality discrepancy. Furthermore, in order to pull together the two types of M-modality images generated from the VIS and IR images in the UMMI space, we propose a distribution consistency loss (DCL) to make the modality distribution of the generated M-modalities images as consistent as possible. Finally, we propose a middle modality network (MMN) to further enhance the discrimination and richness of features in an explicit manner. Extensive experiments have been conducted to validate the superiority of MMN for VI-ReID over some state-of-theart methods on two challenging datasets. The gain of MMN is more than 11.1% and 8.4% in terms of Rank-1 and mAP, respectively, even compared with the latest state-of-the-art methods on the SYSUMM01 dataset.
引用
收藏
页码:788 / 796
页数:9
相关论文
共 50 条
  • [1] Adaptive Middle Modality Alignment Learning for Visible-Infrared Person Re-identification
    Zhang, Yukang
    Yan, Yan
    Lu, Yang
    Wang, Hanzi
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, : 2176 - 2196
  • [2] Towards Grand Unified Representation Learning for Unsupervised Visible-Infrared Person Re-Identification
    Bin Yang
    Chen, Jun
    Ye, Mang
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11035 - 11045
  • [3] Hybrid Modality Metric Learning for Visible-Infrared Person Re-Identification
    Zhang, La
    Guo, Haiyun
    Zhu, Kuan
    Qiao, Honglin
    Huang, Gaopan
    Zhang, Sen
    Zhang, Huichen
    Sun, Jian
    Wang, Jinqiao
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (01)
  • [4] Modality Unifying Network for Visible-Infrared Person Re-Identification
    Yu, Hao
    Cheng, Xu
    Peng, Wei
    Liu, Weihao
    Zhao, Guoying
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 11151 - 11161
  • [5] Learning Modality-Specific Representations for Visible-Infrared Person Re-Identification
    Feng, Zhanxiang
    Lai, Jianhuang
    Xie, Xiaohua
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 579 - 590
  • [6] Modality-agnostic learning for robust visible-infrared person re-identification
    Gong, Shengrong
    Li, Shuomin
    Xie, Gengsheng
    Yao, Yufeng
    Zhong, Shan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (03)
  • [7] Cross-modality consistency learning for visible-infrared person re-identification
    Shao, Jie
    Tang, Lei
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [8] Cross-Modality Semantic Consistency Learning for Visible-Infrared Person Re-Identification
    Liu, Min
    Zhang, Zhu
    Bian, Yuan
    Wang, Xueping
    Sun, Yeqing
    Zhang, Baida
    Wang, Yaonan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 568 - 580
  • [9] Modality Synergy Complement Learning with Cascaded Aggregation for Visible-Infrared Person Re-Identification
    Zhang, Yiyuan
    Zhao, Sanyuan
    Kang, Yuhao
    Shen, Jianbing
    COMPUTER VISION - ECCV 2022, PT XIV, 2022, 13674 : 462 - 479
  • [10] Learning enhancing modality-invariant features for visible-infrared person re-identification
    Zhang, La
    Zhao, Xu
    Du, Haohua
    Sun, Jian
    Wang, Jinqiao
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2025, 16 (01) : 55 - 73