Rank-GCN for Robust Action Recognition

被引:1
作者
Lee, Haetsal [1 ]
Park, Unsang [1 ]
Kim, Ig-Jae [2 ,3 ]
Cho, Junghyun [2 ,3 ]
机构
[1] Sogang Univ, Dept Comp Sci & Engn, Seoul 04107, South Korea
[2] Korea Inst Sci & Technol KIST, Artificial Intelligence & Robot Inst, Seoul 02792, South Korea
[3] Univ Sci & Technol UST, KIST Sch, AI Robot, Seoul 02792, South Korea
关键词
Three-dimensional displays; Spatiotemporal phenomena; Robustness; Feature extraction; Convolutional neural networks; Action recognition; graph convolutional network; dynamic convolutional network;
D O I
10.1109/ACCESS.2022.3202164
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a robust skeleton-based action recognition method with graph convolutional network (GCN) that uses the new adjacency matrix, called Rank-GCN. In Rank-GCN, the biggest change from previous approaches is how the adjacency matrix is generated to accumulate features from neighboring nodes by re-defining "adjacency." The new adjacency matrix, which we call the rank adjacency matrix, is generated by ranking all the nodes according to metrics including the Euclidean distance from the nodes of interest, whereas the previous GCNs methods used only 1-hop neighboring nodes to construct adjacency. By adopting the rank adjacency matrix, we find not only performance improvements but also robustness against swapping, location shifting and dropping of certain nodes. The fact that the human-made rank adjacency matrix wins against the deep-learning-based matrix, implies that there are still some parts that need touch of humans. We expect our Rank-GCN can make performance improvements especially when the predicted human joints are less accurate and unstable.
引用
收藏
页码:91739 / 91749
页数:11
相关论文
共 36 条
[1]   OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields [J].
Cao, Zhe ;
Hidalgo, Gines ;
Simon, Tomas ;
Wei, Shih-En ;
Sheikh, Yaser .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (01) :172-186
[2]  
Carreira J, 2019, Arxiv, DOI arXiv:1907.06987
[3]   Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset [J].
Carreira, Joao ;
Zisserman, Andrew .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4724-4733
[4]   Skeleton-Based Action Recognition with Shift Graph Convolutional Network [J].
Cheng, Ke ;
Zhang, Yifan ;
He, Xiangyu ;
Chen, Weihan ;
Cheng, Jian ;
Lu, Hanqing .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :180-189
[5]   PoTion: Pose MoTion Representation for Action Recognition [J].
Choutas, Vasileios ;
Weinzaepfel, Philippe ;
Revaud, Jerome ;
Schmid, Cordelia .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7024-7033
[6]   Learning Spatiotemporal Features with 3D Convolutional Networks [J].
Du Tran ;
Bourdev, Lubomir ;
Fergus, Rob ;
Torresani, Lorenzo ;
Paluri, Manohar .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :4489-4497
[7]  
Duan Haodong, 2021, ABS210413586 CORR, DOI 10.48550/arXiv.2104.13586
[8]  
Fernando B, 2015, PROC CVPR IEEE, P5378, DOI 10.1109/CVPR.2015.7299176
[9]  
Howard AG, 2017, Arxiv, DOI [arXiv:1704.04861, DOI 10.48550/ARXIV.1704.04861]
[10]   Quo Vadis, Skeleton Action Recognition? [J].
Gupta, Pranay ;
Thatipelli, Anirudh ;
Aggarwal, Aditya ;
Maheshwari, Shubh ;
Trivedi, Neel ;
Das, Sourav ;
Sarvadevabhatla, Ravi Kiran .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (07) :2097-2112