Enhancing skeleton-based action recognition using a knowledge-driven shift graph convolutional network

被引:2
作者
Roy, Ananya [1 ]
Tiwari, Aruna [1 ]
Saurav, Sumeet [2 ]
Singh, Sanjay [2 ]
机构
[1] Indian Inst Technol Indore, Dept Comp Sci & Engn, Khandwa Rd, Indore 453552, Madhya Pradesh, India
[2] CSIR Cent Elect Engn Res Inst CSIR CEERI, Pilani 333031, Rajasthan, India
关键词
Human action recognition; Knowledge graph; Graph convolutional networks; Feature shift; Enhanced feature map;
D O I
10.1016/j.compeleceng.2024.109633
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, there has been a noticeable upsurge in the exploration of human action recognition (HAR) using skeleton data. Compared with video data, skeleton data has advantages, such as lightweightness and high resilience to changes in appearance, lighting conditions, and camera angles. Graph convolutional networks (GCNs) excel in feature extraction from skeletal data, which is non-Euclidean by nature. However, many existing GCN-based methods suffer from high complexities and rigid receptive fields. Consequently, attention has shifted toward the development of lighter architectures with fewer parameters. One such approach is the ShiftGCN, which integrates lightweight feature shift operations to enhance the adaptability of spatiotemporal receptive fields. Despite efficiently capturing distant spatial relationships, this method struggles to differentiate similar actions with subtle variations and requires additional graph connection information. To this end, this study proposes a knowledge-driven shift-GCN (KDS-GCN). The proposed model generates a more detailed and nuanced feature representation by leveraging the integration of graph connectivity knowledge with coordinate information, combined with a lightweight and flexible shift framework, thereby improving action recognition performance. Experiments on four benchmark datasets exhibit the superior performance of the proposed KDS-GCN model while demanding lower computational resources than existing methods.
引用
收藏
页数:16
相关论文
共 56 条
[1]   Graph Convolutional Neural Network for Human Action Recognition: A Comprehensive Survey [J].
Ahmad T. ;
Jin L. ;
Zhang X. ;
Lai S. ;
Tang G. ;
Lin L. .
IEEE Transactions on Artificial Intelligence, 2021, 2 (02) :128-145
[2]   4-Connected Shift Residual Networks [J].
Brown, Andrew ;
Mettes, Pascal ;
Worring, Marcel .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, :1990-1997
[3]   Skeleton-Based Action Recognition With Gated Convolutional Neural Networks [J].
Cao, Congqi ;
Lan, Cuiling ;
Zhang, Yifan ;
Zeng, Wenjun ;
Lu, Hanqing ;
Zhang, Yanning .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (11) :3247-3257
[4]   Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields [J].
Cao, Zhe ;
Simon, Tomas ;
Wei, Shih-En ;
Sheikh, Yaser .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1302-1310
[5]   Channel-wise Topology Refinement Graph Convolution for Skeleton-Based Action Recognition [J].
Chen, Yuxin ;
Zhang, Ziqi ;
Yuan, Chunfeng ;
Li, Bing ;
Deng, Ying ;
Hu, Weiming .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :13339-13348
[6]   Skeleton-Based Action Recognition with Shift Graph Convolutional Network [J].
Cheng, Ke ;
Zhang, Yifan ;
He, Xiangyu ;
Chen, Weihan ;
Cheng, Jian ;
Lu, Hanqing .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :180-189
[7]   The HA4M dataset: Multi-Modal Monitoring of an assembly task for Human Action recognition in Manufacturing [J].
Cicirelli, Grazia ;
Marani, Roberto ;
Romeo, Laura ;
Dominguez, Manuel Garcia ;
Heras, Jonathan ;
Perri, Anna G. G. ;
D'Orazio, Tiziana .
SCIENTIFIC DATA, 2022, 9 (01)
[8]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714
[9]   Efficient Human Action Recognition Interface for Augmented and Virtual Reality Applications Based on Binary Descriptor<bold> </bold> [J].
Fangbemi, Abassin Sourou ;
Liu, Bin ;
Yu, Neng Hai ;
Zhang, Yanxiang .
AUGMENTED REALITY, VIRTUAL REALITY, AND COMPUTER GRAPHICS, AVR 2018, PT I, 2018, 10850 :252-260
[10]  
Fernando B, 2015, PROC CVPR IEEE, P5378, DOI 10.1109/CVPR.2015.7299176