TransGait: Multimodal-based gait recognition with set transformer

被引:30
|
作者
Li, Guodong [1 ]
Guo, Lijun [1 ]
Zhang, Rong [1 ]
Qian, Jiangbo [1 ]
Gao, Shangce [2 ]
机构
[1] NingBo Univ, Fac Elect Engn & Comp Sci, Ningbo, Peoples R China
[2] Univ Toyama, Toyama, Japan
关键词
Gait recognition; Multi-modal; Transformer;
D O I
10.1007/s10489-022-03543-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As a biological feature that can be recognized from a distance, gait has a wide range of applications such as crime prevention, judicial identification, and social security. However, gait recognition is still a challenging task with two problems in the typical gait recognition methods. First, the existing gait recognition methods have weak robustness to the pedestrians' clothing and carryings. Second, the existing temporal modeling methods for gait recognition fail to fully exploit the temporal relationships of the sequence and require that the gait sequence maintain unnecessary sequential constraints. In this paper, we propose a new multi-modal gait recognition framework based on silhouette and pose features to overcome these problems. Joint features of silhouettes and poses provide high discriminability and robustness to the pedestrians' clothing and carryings. Furthermore, we propose a set transformer model with a temporal aggregation operation for obtaining set-level spatio-temporal features. The temporal modeling approach is unaffected by frame permutations and can seamlessly integrate frames from different videos acquired in different scenarios, such as diverse viewing angles. Experiments on two public datasets, CASIA-B and GREW, demonstrate that the proposed method provides state-of-the-art performance. Under the most challenging condition of walking in different clothes on CASIA-B, the proposed method achieves a rank-1 accuracy of 85.8%, outperforming other methods by a significant margin (> 4%).
引用
收藏
页码:1535 / 1547
页数:13
相关论文
共 50 条
  • [1] TransGait: Multimodal-based gait recognition with set transformer
    Guodong Li
    Lijun Guo
    Rong Zhang
    Jiangbo Qian
    Shangce Gao
    Applied Intelligence, 2023, 53 : 1535 - 1547
  • [2] Transformer-Based Multimodal Spatial-Temporal Fusion for Gait Recognition
    Zhang, Jikai
    Ji, Mengyu
    He, Yihao
    Guo, Dongliang
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XV, 2025, 15045 : 494 - 507
  • [3] Gait Recognition by Jointing Transformer and CNN
    Cai, Mingyu
    Wang, Ming
    Zhang, Shunli
    BIOMETRIC RECOGNITION, CCBR 2023, 2023, 14463 : 312 - 321
  • [4] Pedestrian Attribute Recognition Based on Multimodal Transformer
    Liu, Dan
    Song, Wei
    Zhao, Xiaobing
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 422 - 433
  • [5] mmGaitSet: multimodal based gait recognition for countering carrying and clothing changes
    Zhao, Liming
    Guo, Lijun
    Zhang, Rong
    Xie, Xijiong
    Ye, Xulun
    APPLIED INTELLIGENCE, 2022, 52 (02) : 2023 - 2036
  • [6] mmGaitSet: multimodal based gait recognition for countering carrying and clothing changes
    Liming Zhao
    Lijun Guo
    Rong Zhang
    Xijiong Xie
    Xulun Ye
    Applied Intelligence, 2022, 52 : 2023 - 2036
  • [7] Research on Gait Recognition Based on GaitSet and Multimodal Fusion
    Shi, Xiling
    Zhao, Wenqiang
    Pei, Huandou
    Zhai, Hongru
    Gao, Yongxia
    IEEE ACCESS, 2025, 13 : 20017 - 20024
  • [8] TRANSFORMER BASED MULTIMODAL SCENE RECOGNITION IN SOCCER VIDEOS
    Gan, Yaozong
    Togo, Ren
    Ogawa, Takahiro
    Haseyama, Miki
    2022 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (IEEE ICMEW 2022), 2022,
  • [9] Spatial transformer network on skeleton-based gait recognition
    Zhang, Cun
    Chen, Xing-Peng
    Han, Guo-Qiang
    Liu, Xiang-Jie
    EXPERT SYSTEMS, 2023, 40 (06)
  • [10] UGaitNet: Multimodal Gait Recognition With Missing Input Modalities
    Marin-Jimenez, Manuel J.
    Castro, Francisco M.
    Delgado-Escano, Ruben
    Kalogeiton, Vicky
    Guil, Nicolas
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2021, 16 (16) : 5452 - 5462