Spatio-Temporal Fusion Gait Recognition Method Combining Silhouette and Pose

被引:0
作者
Zhang, Chaoyue [1 ]
Zhang, Rong [1 ]
机构
[1] Faculty of Electrical Engineering and Computer Science, Ningbo University, Zhejiang, Ningbo
关键词
attention; gait recognition; multi-modality; multi-scale; occlusion resistance;
D O I
10.3778/j.issn.1002-8331.2204-0500
中图分类号
Q66 [生物力学]; Q811 [仿生学]; Q692 [];
学科分类号
1111 ;
摘要
Most of the existing gait recognition methods are contour-based gait recognition methods, however, contours are easily affected by occlusion, resulting in a decrease in recognition accuracy. In real monitoring scenarios, occlusion is almost inevitable, and improving the accuracy of gait recognition under occlusion is the premise that the algorithm can land in practical applications. Aiming at this problem, a spatio-temporal fusion gait recognition method combining silhouette and pose is proposed. Using the ability of pose to resist occlusion, a multi-modality spatial feature fusion module is designed, and the feature reuse strategy and modal fusion strategy are used to improve the information capacity of spatial features. A multi-scale temporal feature extraction module is designed to extract temporal information at different time scales using independent branches, and an attention-based feature fusion strategy is proposed to integrate temporal information adaptively. A spatial feature set branch is designed to improve the representation of spatial-temporal features in a deeply supervised manner. Experimental results on publicly available datasets show the effectiveness of the proposed method, and the model has good robustness under occlusion. © 2023 Journal of Computer Engineering and Applications Beijing Co., Ltd.; Science Press. All rights reserved.
引用
收藏
页码:135 / 142
页数:7
相关论文
共 18 条
[11]  
ZHU Z, GUO X, YANG T, Et al., Gait recognition in the wild:a benchmark, Proceedings of the 2021 IEEE/ CVF International Conference on Computer Vision, pp. 14789-14799, (2021)
[12]  
WEI S E, RAMAKRISHNA V, KANADE T, Et al., Convolutional pose machines[C], Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, pp. 4724-4732, (2016)
[13]  
TAKEMURA N, MAKIHARA Y, MURAMATSU D, Et al., Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition[J], IPSJ Transactions on Computer Vision & Applications, 10, 1, (2018)
[14]  
HERMANS A, BEYER L, LEIBE B., In defense of the triplet loss for person re-identification, (2017)
[15]  
HUANG G, LIU Z, VAN DER MAATEN L, Et al., Densely connected convolutional networks[C], Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700-4708, (2017)
[16]  
JIE H, LI S, GANG S, Et al., Squeeze-and-excitation networks[C], Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132-7141, (2018)
[17]  
LEE C Y, XIE S, GALLAGHER P, Et al., Deeply-supervised nets[C], Proceedings of the 18th International Conference on Artificial Intelligence and Statistics, pp. 562-570, (2014)
[18]  
JU H, BHANU B., Individual recognition using gait energy image[J], IEEE Transactions on Pattern Analysis and Machine Intelligence, 28, 2, pp. 316-322, (2005)