Enhancing skeleton-based action recognition using a knowledge-driven shift graph convolutional network

被引:2
作者
Roy, Ananya [1 ]
Tiwari, Aruna [1 ]
Saurav, Sumeet [2 ]
Singh, Sanjay [2 ]
机构
[1] Indian Inst Technol Indore, Dept Comp Sci & Engn, Khandwa Rd, Indore 453552, Madhya Pradesh, India
[2] CSIR Cent Elect Engn Res Inst CSIR CEERI, Pilani 333031, Rajasthan, India
关键词
Human action recognition; Knowledge graph; Graph convolutional networks; Feature shift; Enhanced feature map;
D O I
10.1016/j.compeleceng.2024.109633
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, there has been a noticeable upsurge in the exploration of human action recognition (HAR) using skeleton data. Compared with video data, skeleton data has advantages, such as lightweightness and high resilience to changes in appearance, lighting conditions, and camera angles. Graph convolutional networks (GCNs) excel in feature extraction from skeletal data, which is non-Euclidean by nature. However, many existing GCN-based methods suffer from high complexities and rigid receptive fields. Consequently, attention has shifted toward the development of lighter architectures with fewer parameters. One such approach is the ShiftGCN, which integrates lightweight feature shift operations to enhance the adaptability of spatiotemporal receptive fields. Despite efficiently capturing distant spatial relationships, this method struggles to differentiate similar actions with subtle variations and requires additional graph connection information. To this end, this study proposes a knowledge-driven shift-GCN (KDS-GCN). The proposed model generates a more detailed and nuanced feature representation by leveraging the integration of graph connectivity knowledge with coordinate information, combined with a lightweight and flexible shift framework, thereby improving action recognition performance. Experiments on four benchmark datasets exhibit the superior performance of the proposed KDS-GCN model while demanding lower computational resources than existing methods.
引用
收藏
页数:16
相关论文
共 56 条
[31]   Enhanced skeleton visualization for view invariant human action recognition [J].
Liu, Mengyuan ;
Liu, Hong ;
Chen, Chen .
PATTERN RECOGNITION, 2017, 68 :346-362
[32]   HYPERBOLIC SPATIAL TEMPORAL GRAPH CONVOLUTIONAL NETWORKS [J].
Mostafa, Abdelrahman ;
Peng, Wei ;
Zhao, Guoying .
2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, :3301-3305
[33]  
Kipf TN, 2017, Arxiv, DOI arXiv:1609.02907
[34]   A survey on video-based Human Action Recognition: recent updates, datasets, challenges, and applications [J].
Pareek, Preksha ;
Thakkar, Ankit .
ARTIFICIAL INTELLIGENCE REVIEW, 2021, 54 (03) :2259-2322
[35]  
Peng W, 2020, 34 AAAI C ART INT, P23
[36]   Rethinking the ST-GCNs for 3D skeleton-based human action recognition [J].
Peng, Wei ;
Shi, Jingang ;
Varanka, Tuomas ;
Zhao, Guoying .
NEUROCOMPUTING, 2021, 454 :45-53
[37]   Improved skeleton-based activity recognition using convolutional block attention module [J].
Qin, Jing ;
Zhang, Shugang ;
Wang, Yiguo ;
Yang, Fei ;
Zhong, Xin ;
Lu, Weigang .
COMPUTERS & ELECTRICAL ENGINEERING, 2024, 116
[38]   NTU RGB plus D: A Large Scale Dataset for 3D Human Activity Analysis [J].
Shahroudy, Amir ;
Liu, Jun ;
Ng, Tian-Tsong ;
Wang, Gang .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1010-1019
[39]   Multiscale 3D-Shift Graph Convolution Network for Emotion Recognition From Human Actions [J].
Shi, Henglin ;
Peng, Wei ;
Chen, Haoyu ;
Liu, Xin ;
Zhao, Guoying .
IEEE INTELLIGENT SYSTEMS, 2022, 37 (04) :103-110
[40]   Skeleton-Based Action Recognition With Multi-Stream Adaptive Graph Convolutional Networks [J].
Shi, Lei ;
Zhang, Yifan ;
Cheng, Jian ;
Lu, Hanqing .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :9532-9545