Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning

被引:277
|
作者
Ali, Saad [1 ]
Shah, Mubarak [2 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[2] Univ Cent Florida, Comp Vis Lab, Sch Elect Engn & Comp Sci, Harris Corp Engn Ctr, Orlando, FL 32816 USA
关键词
Action recognition; motion; video analysis; principal component analysis; kinematic features;
D O I
10.1109/TPAMI.2008.284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a set of kinematic features that are derived from the optical flow for human action recognition in videos. The set of kinematic features includes divergence, vorticity, symmetric and antisymmetric flow fields, second and third principal invariants of flow gradient and rate of strain tensor, and third principal invariant of rate of rotation tensor. Each kinematic feature, when computed from the optical flow of a sequence of images, gives rise to a spatiotemporal pattern. It is then assumed that the representative dynamics of the optical flow are captured by these spatiotemporal patterns in the form of dominant kinematic trends or kinematic modes. These kinematic modes are computed by performing Principal Component Analysis (PCA) on the spatiotemporal volumes of the kinematic features. For classification, we propose the use of multiple instance learning (MIL) in which each action video is represented by a bag of kinematic modes. Each video is then embedded into a kinematic-mode-based feature space and the coordinates of the video in that space are used for classification using the nearest neighbor algorithm. The qualitative and quantitative results are reported on the benchmark data sets.
引用
收藏
页码:288 / 303
页数:16
相关论文
共 50 条
  • [21] Learning Spatiotemporal-Selected Representations in Videos for Action Recognition
    Zhang, Jiachao
    Tong, Ying
    Jiao, Liangbao
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2023, 32 (12)
  • [22] Multilayer deep features with multiple kernel learning for action recognition
    Sheng, Biyun
    Li, Jun
    Xiao, Fu
    Yang, Wankou
    NEUROCOMPUTING, 2020, 399 : 65 - 74
  • [23] Boosting VLAD with Double Assignment using Deep Features for Action Recognition in Videos
    Duta, Ionut C.
    Nguyen, Tuan A.
    Aizawa, Kiyoharu
    Ionescu, Bogdan
    Sebe, Nicu
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 2210 - 2215
  • [24] Study on human Action Recognition Algorithms in videos
    He Chun-Lin
    Pan Wei
    PROCEEDINGS OF THE 2015 INTERNATIONAL SYMPOSIUM ON COMPUTERS & INFORMATICS, 2015, 13 : 703 - 709
  • [25] Action Recognition Using Multiple Pooling Strategies of CNN Features
    Hu, Haifeng
    Liao, Zhongke
    Xiao, Xiang
    NEURAL PROCESSING LETTERS, 2019, 50 (01) : 379 - 396
  • [26] Action Recognition Using Multiple Pooling Strategies of CNN Features
    Haifeng Hu
    Zhongke Liao
    Xiang Xiao
    Neural Processing Letters, 2019, 50 : 379 - 396
  • [27] Multimodal Multipart Learning for Action Recognition in Depth Videos
    Shahroudy, Amir
    Ng, Tian-Tsong
    Yang, Qingxiong
    Wang, Gang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (10) : 2123 - 2129
  • [28] Comparison of Cuboid and Tracklet Features for Action Recognition on Surveillance Videos
    Bayram, Ulya
    Ulusoy, Ilkay
    Cicekli, Nihan Kesim
    2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013,
  • [29] Multiple stream deep learning model for human action recognition
    Gu, Ye
    Ye, Xiaofeng
    Sheng, Weihua
    Ou, Yongsheng
    Li, Yongqiang
    IMAGE AND VISION COMPUTING, 2020, 93
  • [30] A Spatio-Temporal Deep Learning Approach For Human Action Recognition in Infrared Videos
    Shah, Anuj K.
    Ghosh, Ripul
    Akula, Aparna
    OPTICS AND PHOTONICS FOR INFORMATION PROCESSING XII, 2018, 10751