Action recognition and tracking via deep representation extraction and motion bases learning

被引:0
|
作者
Hao-Ting Li
Yung-Pin Liu
Yun-Kai Chang
Chen-Kuo Chiang
机构
[1] National Chung Cheng University,Department of Computer Science and Information Engineering, Advanced Institute of Manufacturingwith High
来源
Multimedia Tools and Applications | 2022年 / 81卷
关键词
Action recognition; Motion bases; Action tracking; Deep learning;
D O I
暂无
中图分类号
学科分类号
摘要
Action recognition and positional tracking are critical issues in many applications in Virtual Reality (VR). In this paper, a novel feature representation method is proposed to recognize actions based on sensor signals. The feature extraction is achieved by jointly learning Convolutional Auto-Encoder (CAE) and the representation of motion bases via clustering, which is called the Sequence of Cluster Centroids (SoCC). Then, the learned features are used to train the action recognition classifier. We have collected new dataset of actions of limbs by sensor signals. In addition, a novel action tracking method is proposed for the VR environment. It extends the sensor signals from three Degrees of Freedom (DoF) of rotation to 6DoF of position plus rotation. Experimental results demonstrate that CAE-SoCC feature is effective for action recognition and accurate prediction of position displacement.
引用
收藏
页码:11845 / 11864
页数:19
相关论文
共 50 条
  • [1] Action recognition and tracking via deep representation extraction and motion bases learning
    Li, Hao-Ting
    Liu, Yung-Pin
    Chang, Yun-Kai
    Chiang, Chen-Kuo
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (09) : 11845 - 11864
  • [2] GMNet: an action recognition network with global motion representation
    Mingwei Liu
    Yi Zhang
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 1683 - 1693
  • [3] GMNet: an action recognition network with global motion representation
    Liu, Mingwei
    Zhang, Yi
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (05) : 1683 - 1693
  • [4] Face Recognition via Deep Learning and Constraint Sparse Representation
    Zhang J.-W.
    Niu S.-Z.
    Cao Z.-Y.
    Wang X.-Y.
    Beijing Ligong Daxue Xuebao/Transaction of Beijing Institute of Technology, 2019, 39 (03): : 255 - 261
  • [5] Learning hierarchical video representation for action recognition
    Li Q.
    Qiu Z.
    Yao T.
    Mei T.
    Rui Y.
    Luo J.
    International Journal of Multimedia Information Retrieval, 2017, 6 (1) : 85 - 98
  • [6] Deep Learning for Human Action Recognition
    Shekokar, R. U.
    Kale, S. N.
    2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,
  • [7] Deep Video Understanding: Representation Learning, Action Recognition, and Language Generation
    Mei, Tao
    PROCEEDINGS OF THE 1ST WORKSHOP AND CHALLENGE ON COMPREHENSIVE VIDEO UNDERSTANDING IN THE WILD (COVIEW'18), 2018, : 1 - 1
  • [8] A spatiotemporal and motion information extraction network for action recognition
    Wang, Wei
    Wang, Xianmin
    Zhou, Mingliang
    Wei, Xuekai
    Li, Jing
    Ren, Xiaojun
    Zong, Xuemei
    WIRELESS NETWORKS, 2024, 30 (06) : 5389 - 5405
  • [9] Action Recognition by Learning Deep Multi-Granular Spatio-Temporal Video Representation
    Li, Qing
    Qiu, Zhaofan
    Yao, Ting
    Mei, Tao
    Rui, Yong
    Luo, Jiebo
    ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 159 - 166
  • [10] Learning Self-Correlation in Space and Time as Motion Representation for Action Recognition
    Zhang, Yi
    Li, Yuchang
    Liu, Mingwei
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 1747 - 1751