Algorithm for Skeleton Action Recognition by Integrating Attention Mechanism and Convolutional Neural Networks

被引:0
作者
Liu, Jianhua [1 ]
机构
[1] Weifang Univ, Coll Phys Educ, Weifang 261061, Peoples R China
关键词
Attention mechanism; convolutional neural network; action recognition; central differential network; spacetime converter; directed graph convolution;
D O I
10.14569/IJACSA.2023.0140867
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
An action recognition model based on 3D skeleton data may experience a decrease in recognition accuracy when facing complex backgrounds, and it is easy to overlook the local connection between dynamic gradient information and dynamic actions, resulting in a decrease in the fault tolerance of the constructed model. To achieve accurate and fast capture of human skeletal movements, a directed graph convolutional network recognition model that integrates attention mechanism and convolutional neural network is proposed. By combining spacetime converter and central differential graph convolution, a corresponding central differential converter graph convolutional network model is constructed to obtain dynamic gradient information in actions and calculate local connections between dynamic actions. The research outcomes express that the cross-target benchmark recognition rate of the directed graph convolutional network recognition model is 92.3%, and the cross-view benchmark recognition rate is 97.3%. The accuracy of Top -1 is 37.6%, and the accuracy of Top-5 is 60.5%. The cross-target recognition rate of the central differential converter graph convolutional network model is 92.9%, and the cross-view benchmark recognition rate is 97.5%. Undercross-target and cross-view benchmarks, the average recognition accuracy for similar actions is 81.3% and 88.9%, respectively. The accuracy of the entire action recognition model in single-person multi-person action recognition experiments is 95.0%. The outcomes denote that the model constructed by the research institute has higher recognition rate and more stable performance compared to existing neural network recognition models, and has certain research value.
引用
收藏
页码:604 / 613
页数:10
相关论文
共 19 条
[11]   AIR-Act2Act: Human-human interaction dataset for teaching non-verbal social behaviors to robots [J].
Ko, Woo-Ri ;
Jang, Minsu ;
Lee, Jaeyeon ;
Kim, Jaehong .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2021, 40 (4-5) :691-697
[12]  
Lei Y., 2022, J. Comput. Cogn. Eng, V1, P83
[13]   Three-Dimensional Diffusion Model in Sports Dance Video Human Skeleton Detection and Extraction [J].
Li, Zhi .
ADVANCES IN MATHEMATICAL PHYSICS, 2021, 2021
[14]   Action recognition for sports video analysis using part-attention spatio-temporal graph convolutional network [J].
Liu, Jiatong ;
Che, Yanli .
JOURNAL OF ELECTRONIC IMAGING, 2021, 30 (03)
[15]   Graph transformer network with temporal kernel attention for skeleton-based action recognition [J].
Liu, Yanan ;
Zhang, Hao ;
Xu, Dan ;
He, Kangjian .
KNOWLEDGE-BASED SYSTEMS, 2022, 240
[16]   Human Motion Gesture Recognition Based on Computer Vision [J].
Ma, Rui ;
Zhang, Zhendong ;
Chen, Enqing .
COMPLEXITY, 2021, 2021
[17]   Motion boundary emphasised optical flow method for human action recognition [J].
Peng, Cheng ;
Huang, Haozhi ;
Tsoi, Ah-Chung ;
Lo, Sio-Long ;
Liu, Yun ;
Yang, Zi-yi .
IET COMPUTER VISION, 2020, 14 (06) :378-390
[18]  
Yang Y., 2022, J Comput Cognit Eng, V1, P32, DOI [DOI 10.47852/BONVIEWJCCE19919, 10.47852/bonviewjcce19919.[8]Q, 10.47852/bonviewJCCE19919]
[19]   Multi-mode neural network for human action recognition [J].
Zhao, Haohua ;
Xue, Weichen ;
Li, Xiaobo ;
Gu, Zhangxuan ;
Niu, Li ;
Zhang, Liqing .
IET COMPUTER VISION, 2020, 14 (08) :587-596