Skeleton-Based Action Recognition with Directed Graph Neural Networks

被引:574
|
作者
Shi, Lei [1 ,2 ]
Zhang, Yifan [1 ,2 ]
Cheng, Jian [1 ,2 ,3 ]
Lu, Hanqing [1 ,2 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
[3] CAS Ctr Excellence Brain Sci & Intelligence Techn, Beijing, Peoples R China
来源
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019) | 2019年
基金
中国国家自然科学基金;
关键词
POSE;
D O I
10.1109/CVPR.2019.00810
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The skeleton data have been widely used for the action recognition tasks since they can robustly accommodate dynamic circumstancesand complex backgrounds. In existing methods, both the joint and bone information in skeleton data have been proved to be of great help for action recognition tasks. However, how to incorporate these two types of data to best take advantage of the relationshipbetween joints and bones remains a problem to be solved. In this work, we represent the skeleton data as a directed acyclic graph (DAG) based on the kinematic dependency between the joints and bones in the naturalhuman body. A novel directedgraphneural network is designedspecially to extract the information ofjoints, bones and their relationshipsand make prediction based on the extractedfeatures. In addition, to betterfit the action recognitiontask, the topological structure of the graph is made adaptivebased on the training process, which brings notable improvement. Moreover, the motion information of the skeleton sequence is exploited and combined with the spatial information to further enhance the performance in a two-stream framework. Ourfinal model is tested on two large-scaledatasets, NTU-RGBD and Skeleton-Kinetics, and exceeds state-of-the-artperformance on both of them.
引用
收藏
页码:7904 / 7913
页数:10
相关论文
共 50 条
  • [1] Graph Edge Convolutional Neural Networks for Skeleton-Based Action Recognition
    Zhang, Xikun
    Xu, Chang
    Tian, Xinmei
    Tao, Dacheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (08) : 3047 - 3060
  • [2] SKELETON-BASED ACTION RECOGNITION WITH CONVOLUTIONAL NEURAL NETWORKS
    Li, Chao
    Zhong, Qiaoyong
    Xie, Di
    Pu, Shiliang
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2017,
  • [3] Selective directed graph convolutional network for skeleton-based action recognition
    Ke, Chengyuan
    Liu, Sheng
    Feng, Yuan
    Chen, Shengyong
    PATTERN RECOGNITION LETTERS, 2025, 190 : 141 - 146
  • [4] Recurrent graph convolutional networks for skeleton-based action recognition
    Zhu, Guangming
    Yang, Lu
    Zhang, Liang
    Shen, Peiyi
    Song, Juan
    Proceedings - International Conference on Pattern Recognition, 2020, : 1352 - 1359
  • [5] Recurrent Graph Convolutional Networks for Skeleton-based Action Recognition
    Zhu, Guangming
    Yang, Lu
    Zhang, Liang
    Shen, Peiyi
    Song, Juan
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 1352 - 1359
  • [6] Fourier analysis on robustness of graph convolutional neural networks for skeleton-based action recognition
    Tanaka, Nariki
    Kera, Hiroshi
    Kawamoto, Kazuhiko
    Computer Vision and Image Understanding, 2024, 240
  • [7] Spatial-temporal graph neural ODE networks for skeleton-based action recognition
    Pan, Longji
    Lu, Jianguang
    Tang, Xianghong
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [8] Fourier analysis on robustness of graph convolutional neural networks for skeleton-based action recognition
    Tanaka, Nariki
    Kera, Hiroshi
    Kawamoto, Kazuhiko
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 240
  • [9] Spatial-temporal graph neural ODE networks for skeleton-based action recognition
    Longji Pan
    Jianguang Lu
    Xianghong Tang
    Scientific Reports, 14
  • [10] Skeleton-Based Action Recognition With Gated Convolutional Neural Networks
    Cao, Congqi
    Lan, Cuiling
    Zhang, Yifan
    Zeng, Wenjun
    Lu, Hanqing
    Zhang, Yanning
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2019, 29 (11) : 3247 - 3257