Transformer-based neural architecture search for effective visible-infrared person re-identification

被引:0
|
作者
Sarker, Prodip Kumar [1 ]
机构
[1] Begum Rokeya Univ, Dept Comp Sci & Engn, Rangpur 5400, Bangladesh
关键词
Transformer; Neural architecture search; Attention mechanism; Feature extraction; Cross-modality;
D O I
10.1016/j.neucom.2024.129257
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visible-infrared person re-identification (VI-reID) is a complex task insecurity and video surveillance that aims to identify and match a person captured by various non-overlapping cameras. In recent years, there has been a notable advancement in reID owing to the development of transformer-based architectures. Although many existing methods emphasize on learning both modality-specific and shared features, challenges remain in fully exploiting the complementary information between infrared and visible modalities. Consequently, there is still opportunity to increase retrieval performance by effectively comprehending and integrating cross- modality semantic information. These designs often have problems with model complexity and time-consuming processes. To tackle these issues, we employ a novel transformer-based neural architecture search (TNAS) deep learning approach for effective VI-reID. To alleviate modality gaps, we first introduce a global-local transformer (GLT) module that captures features at both global and local levels across different modalities, contributing to better feature representation and matching. Then, an efficient neural architecture search (NAS) module is developed to search for the optimal transformer-based architecture, which further enhances the performance of VI-reID. Additionally, we introduce distillation loss and modality discriminative (MD) loss to examine the potential consistency between different modalities to promote intermodality separation between classes and intramodality compactness within classes. Experimental results on two challenging benchmark datasets illustrate that our developed model achieves state-of-the-art results, outperforming existing VI-reID methods.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Auxiliary Representation Guided Network for Visible-Infrared Person Re-Identification
    Qi, Mengzan
    Chan, Sixian
    Hang, Chen
    Zhang, Guixu
    Zeng, Tieyong
    Li, Zhi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 340 - 355
  • [22] Contrastive Learning with Information Compensation for Visible-Infrared Person Re-Identification
    Zhang, La
    Guo, Haiyun
    Zhao, Xu
    Sun, Jian
    Wang, Jinqiao
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 1266 - 1271
  • [23] Visible-Infrared Person Re-Identification Via Feature Constrained Learning
    Zhang Jing
    Chen Guangfeng
    LASER & OPTOELECTRONICS PROGRESS, 2024, 61 (12)
  • [24] A PERSON RE-IDENTIFICATION BASELINE BASED ON ATTENTION BLOCK NEURAL ARCHITECTURE SEARCH
    Sun, Jia
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 841 - 845
  • [25] Dynamic Center Aggregation Loss With Mixed Modality for Visible-Infrared Person Re-Identification
    Kong, Jun
    He, Qibin
    Jiang, Min
    Liu, Tianshan
    IEEE SIGNAL PROCESSING LETTERS, 2021, 28 : 2003 - 2007
  • [26] A Generative-Based Image Fusion Strategy for Visible-Infrared Person Re-Identification
    Qi, Jia
    Liang, Tengfei
    Liu, Wu
    Li, Yidong
    Jin, Yi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) : 518 - 533
  • [27] Visible-infrared person re-identification model based on feature consistency and modal indistinguishability
    Jia Sun
    Yanfeng Li
    Houjin Chen
    Yahui Peng
    Jinlei Zhu
    Machine Vision and Applications, 2023, 34
  • [28] Visible-infrared person re-identification model based on feature consistency and modal indistinguishability
    Sun, Jia
    Li, Yanfeng
    Chen, Houjin
    Peng, Yahui
    Zhu, Jinlei
    MACHINE VISION AND APPLICATIONS, 2023, 34 (01)
  • [29] Dual-attentive cascade clustering learning for visible-infrared person re-identification
    Wang, Xianju
    Chen, Cuiqun
    Zhu, Yong
    Chen, Shuguang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (07) : 19729 - 19746
  • [30] DMANet: Dual-modality alignment network for visible-infrared person re-identification
    Cheng, Xu
    Deng, Shuya
    Yu, Hao
    Zhao, Guoying
    PATTERN RECOGNITION, 2025, 157