Enhanced Adjacency Matrix-Based Lightweight Graph Convolution Network for Action Recognition

被引:5
作者
Zhang, Daqing [1 ]
Deng, Hongmin [1 ]
Zhi, Yong [1 ]
机构
[1] Sichuan Univ, Sch Elect & Informat Engn, Chengdu 610064, Peoples R China
基金
中国国家自然科学基金;
关键词
action recognition; skeleton data; CA-EAMGCN; feature selection; combinatorial attention; MOTION;
D O I
10.3390/s23146397
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Graph convolutional networks (GCNs), which extend convolutional neural networks (CNNs) to non-Euclidean structures, have been utilized to promote skeleton-based human action recognition research and have made substantial progress in doing so. However, there are still some challenges in the construction of recognition models based on GCNs. In this paper, we propose an enhanced adjacency matrix-based graph convolutional network with a combinatorial attention mechanism (CA-EAMGCN) for skeleton-based action recognition. Firstly, an enhanced adjacency matrix is constructed to expand the model's perceptive field of global node features. Secondly, a feature selection fusion module (FSFM) is designed to provide an optimal fusion ratio for multiple input features of the model. Finally, a combinatorial attention mechanism is devised. Specifically, our spatial-temporal (ST) attention module and limb attention module (LAM) are integrated into a multi-input branch and a mainstream network of the proposed model, respectively. Extensive experiments on three large-scale datasets, namely the NTU RGB+D 60, NTU RGB+D 120 and UAV-Human datasets, show that the proposed model takes into account both requirements of light weight and recognition accuracy. This demonstrates the effectiveness of our method.
引用
收藏
页数:20
相关论文
共 41 条
[1]   Human Action Recognition: Pose-based Attention draws focus to Hands [J].
Baradel, Fabien ;
Wolf, Christian ;
Mille, Julien .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, :604-613
[2]   R-Net: A novel fully convolutional network-based infrared image segmentation method for intelligent human behavior analysis [J].
Chen, Shaohui ;
Xu, Xiaogang ;
Yang, Ningyu ;
Chen, Xianghua ;
Du, Feng ;
Ding, Shuyong ;
Gao, Wei .
INFRARED PHYSICS & TECHNOLOGY, 2022, 123
[3]   Action recognition of individuals on an airport apron based on tracking bounding boxes of the thermal infrared target [J].
Ding, Meng ;
Ding, Yuan-yuan ;
Wu, Xiao-zhou ;
Wang, Xu-hui ;
Xu, Yu-bin .
INFRARED PHYSICS & TECHNOLOGY, 2021, 117
[4]  
Du Y, 2015, PROC CVPR IEEE, P1110, DOI 10.1109/CVPR.2015.7298714
[5]  
Fernando B, 2015, PROC CVPR IEEE, P5378, DOI 10.1109/CVPR.2015.7299176
[6]   A survey on visual surveillance of object motion and behaviors [J].
Hu, WM ;
Tan, TN ;
Wang, L ;
Maybank, S .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2004, 34 (03) :334-352
[7]   VISUAL-PERCEPTION OF BIOLOGICAL MOTION AND A MODEL FOR ITS ANALYSIS [J].
JOHANSSON, G .
PERCEPTION & PSYCHOPHYSICS, 1973, 14 (02) :201-211
[8]   AGCNNs: Attention-guided convolutional neural networks for infrared head pose estimation in assisted driving system [J].
Ju, Jianping ;
Zheng, Hong ;
Li, Congcong ;
Li, Xi ;
Liu, Hai ;
Liu, Tingting .
INFRARED PHYSICS & TECHNOLOGY, 2022, 123
[9]   Interpretable 3D Human Action Analysis with Temporal Convolutional Networks [J].
Kim, Tae Soo ;
Reiter, Austin .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :1623-1631
[10]   Tower Crane remote wireless monitoring system based on Modbus/TCP protocol [J].
Li, Bo ;
Chen, Geng ;
Wang, Le ;
Hao, Zhe .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND ENGINEERING (CSE) AND IEEE/IFIP INTERNATIONAL CONFERENCE ON EMBEDDED AND UBIQUITOUS COMPUTING (EUC), VOL 2, 2017, :187-190