Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning

被引:277
|
作者
Ali, Saad [1 ]
Shah, Mubarak [2 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[2] Univ Cent Florida, Comp Vis Lab, Sch Elect Engn & Comp Sci, Harris Corp Engn Ctr, Orlando, FL 32816 USA
关键词
Action recognition; motion; video analysis; principal component analysis; kinematic features;
D O I
10.1109/TPAMI.2008.284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a set of kinematic features that are derived from the optical flow for human action recognition in videos. The set of kinematic features includes divergence, vorticity, symmetric and antisymmetric flow fields, second and third principal invariants of flow gradient and rate of strain tensor, and third principal invariant of rate of rotation tensor. Each kinematic feature, when computed from the optical flow of a sequence of images, gives rise to a spatiotemporal pattern. It is then assumed that the representative dynamics of the optical flow are captured by these spatiotemporal patterns in the form of dominant kinematic trends or kinematic modes. These kinematic modes are computed by performing Principal Component Analysis (PCA) on the spatiotemporal volumes of the kinematic features. For classification, we propose the use of multiple instance learning (MIL) in which each action video is represented by a bag of kinematic modes. Each video is then embedded into a kinematic-mode-based feature space and the coordinates of the video in that space are used for classification using the nearest neighbor algorithm. The qualitative and quantitative results are reported on the benchmark data sets.
引用
收藏
页码:288 / 303
页数:16
相关论文
共 50 条
  • [31] Human action recognition with salient trajectories and multiple kernel learning
    Yang Yi
    Pan Hu
    Xiaokang Deng
    Multimedia Tools and Applications, 2018, 77 : 17709 - 17730
  • [32] Human action recognition in complex live videos using graph convolutional network*
    Bharathi, A.
    Sridevi, M.
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 110
  • [33] HUMAN ACTIVITY DETECTION AND ACTION RECOGNITION IN VIDEOS USING CONVOLUTIONAL NEURAL NETWORKS
    Basavaiah, Jagadeesh
    Patil, Chandrashekar Mohan
    JOURNAL OF INFORMATION AND COMMUNICATION TECHNOLOGY-MALAYSIA, 2020, 19 (02): : 157 - 183
  • [34] Human action recognition with salient trajectories and multiple kernel learning
    Yi, Yang
    Hu, Pan
    Deng, Xiaokang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (14) : 17709 - 17730
  • [35] DMMs-Based Multiple Features Fusion for Human Action Recognition
    Bulbul, Mohammad Farhad
    Jiang, Yunsheng
    Ma, Jinwen
    INTERNATIONAL JOURNAL OF MULTIMEDIA DATA ENGINEERING & MANAGEMENT, 2015, 6 (04) : 23 - 39
  • [36] Dynamic Hand Gesture Recognition Using Kinematic Features Based on Dynamic Time Warping
    Pang, Haibo
    Ding, Youdong
    ADVANCES IN COMPUTING, CONTROL AND INDUSTRIAL ENGINEERING, 2012, 235 : 68 - +
  • [37] Human activity recognition in videos using a single example
    Roshtkhari, Mehrsan Javan
    Levine, Martin D.
    IMAGE AND VISION COMPUTING, 2013, 31 (11) : 864 - 876
  • [38] Jointly Learning Multi-view Features for Human Action Recognition
    Wang, Ruoshi
    Liu, Zhigang
    Yin, Ziyang
    PROCEEDINGS OF THE 32ND 2020 CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2020), 2020, : 4858 - 4861
  • [39] Learning features combination for human action recognition from skeleton sequences
    Luvizon, Diogo Carbonera
    Tabia, Hedi
    Picard, David
    PATTERN RECOGNITION LETTERS, 2017, 99 : 13 - 20
  • [40] Action-Scene Model for Human Action Recognition from Videos*
    Zhang, Yifei
    Qu, Wen
    Wang, Daling
    2ND AASRI CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND BIOINFORMATICS, 2014, 6 : 111 - 117