A Novel Action Recognition Scheme Based on Spatial-Temporal Pyramid Model

被引:0
|
作者
Zhao, Hengying [1 ]
Xiang, Xinguang [1 ]
机构
[1] Nanjing Univ Sci & Technol, Sch Comp Sci & Engn, Nanjing 210094, Peoples R China
来源
ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II | 2018年 / 10736卷
关键词
Action recognition; Spatial-temporal; Multi-scale; Visual dictionary; DENSE;
D O I
10.1007/978-3-319-77383-4_21
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recognizing actions is one of the most important challenges in computer vision. In this paper, we propose a novel action recognition scheme based on spatial-temporal pyramid model. Firstly, we extract the basic visual feature descriptors for each video. Secondly, we construct visual dictionary on the whole visual features set. Thirdly, we construct a novel spatial-temporal pyramid model by dividing the visual features set of each video into multi-scale blocks in 2-dimensional space domain and 1-dimensional time domain separately. Then we calculate the distribution histogram representation for each block of different scales by using the bag-of-features model and our new visual dictionary. At last, we normalize the final descriptors for videos and then recognize the actions using SVM. Experimental results show that our scheme achieves more accurate for action recognition compared with several state-of-the-art methods.
引用
收藏
页码:212 / 221
页数:10
相关论文
共 50 条
  • [41] Two-stream spatial-temporal neural networks for pose-based action recognition
    Wang, Zixuan
    Zhu, Aichun
    Hu, Fangqiang
    Wu, Qianyu
    Li, Yifeng
    JOURNAL OF ELECTRONIC IMAGING, 2020, 29 (04)
  • [42] MSST-RT: Multi-Stream Spatial-Temporal Relative Transformer for Skeleton-Based Action Recognition
    Sun, Yan
    Shen, Yixin
    Ma, Liyan
    SENSORS, 2021, 21 (16)
  • [43] Exploring a rich spatial-temporal dependent relational model for skeleton-based action recognition by bidirectional LSTM-CNN
    Zhu, Aichun
    Wu, Qianyu
    Cui, Ran
    Wang, Tian
    Hang, Wenlong
    Hua, Gang
    Snoussi, Hichem
    NEUROCOMPUTING, 2020, 414 : 90 - 100
  • [44] Action Recognition by Fusing Spatial-Temporal Appearance and The Local Distribution of Interest Points
    Lu, Mengmeng
    Zhang, Liang
    PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON FUTURE COMPUTER AND COMMUNICATION ENGINEERING, 2014, 111 : 75 - 78
  • [45] Bi-direction hierarchical LSTM with spatial-temporal attention for action recognition
    Yang, Haodong
    Zhang, Jun
    Li, Shuohao
    Luo, Tingjin
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 36 (01) : 775 - 786
  • [46] R-STAN: Residual Spatial-Temporal Attention Network for Action Recognition
    Liu, Quanle
    Che, Xiangjiu
    Bie, Mei
    IEEE ACCESS, 2019, 7 : 82246 - 82255
  • [47] Human Action Recognition by Fusion of Convolutional Neural Networks and spatial-temporal Information
    Li, Weisheng
    Ding, Yahui
    8TH INTERNATIONAL CONFERENCE ON INTERNET MULTIMEDIA COMPUTING AND SERVICE (ICIMCS2016), 2016, : 255 - 259
  • [48] Learning Semantic-Aware Spatial-Temporal Attention for Interpretable Action Recognition
    Fu, Jie
    Gao, Junyu
    Xu, Changsheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (08) : 5213 - 5224
  • [49] STA-CNN: Convolutional Spatial-Temporal Attention Learning for Action Recognition
    Yang, Hao
    Yuan, Chunfeng
    Zhang, Li
    Sun, Yunda
    Hu, Weiming
    Maybank, Stephen J.
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 5783 - 5793
  • [50] Attention module-based spatial-temporal graph convolutional networks for skeleton-based action recognition
    Kong, Yinghui
    Li, Li
    Zhang, Ke
    Ni, Qiang
    Han, Jungong
    JOURNAL OF ELECTRONIC IMAGING, 2019, 28 (04)