Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning

被引:277
|
作者
Ali, Saad [1 ]
Shah, Mubarak [2 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[2] Univ Cent Florida, Comp Vis Lab, Sch Elect Engn & Comp Sci, Harris Corp Engn Ctr, Orlando, FL 32816 USA
关键词
Action recognition; motion; video analysis; principal component analysis; kinematic features;
D O I
10.1109/TPAMI.2008.284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a set of kinematic features that are derived from the optical flow for human action recognition in videos. The set of kinematic features includes divergence, vorticity, symmetric and antisymmetric flow fields, second and third principal invariants of flow gradient and rate of strain tensor, and third principal invariant of rate of rotation tensor. Each kinematic feature, when computed from the optical flow of a sequence of images, gives rise to a spatiotemporal pattern. It is then assumed that the representative dynamics of the optical flow are captured by these spatiotemporal patterns in the form of dominant kinematic trends or kinematic modes. These kinematic modes are computed by performing Principal Component Analysis (PCA) on the spatiotemporal volumes of the kinematic features. For classification, we propose the use of multiple instance learning (MIL) in which each action video is represented by a bag of kinematic modes. Each video is then embedded into a kinematic-mode-based feature space and the coordinates of the video in that space are used for classification using the nearest neighbor algorithm. The qualitative and quantitative results are reported on the benchmark data sets.
引用
收藏
页码:288 / 303
页数:16
相关论文
共 50 条
  • [1] A Novel Dictionary Learning based Multiple Instance Learning Approach to Action Recognition from Videos
    Roy, Abhinaba
    Banerjee, Biplab
    Murino, Vittorio
    ICPRAM: PROCEEDINGS OF THE 6TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS, 2017, : 519 - 526
  • [2] Human action recognition with graph-based multiple-instance learning
    Yi, Yang
    Lin, Maoqing
    PATTERN RECOGNITION, 2016, 53 : 148 - 162
  • [3] Kinematic Features For Human Action Recognition Using Restricted Boltzmann Machines
    Arinaldi, Ahmad
    Fanany, Mohamad Ivan
    2016 4TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGY (ICOICT), 2016,
  • [4] Human Action Recognition in Videos Using Hybrid Motion Features
    Liu, Si
    Liu, Jing
    Zhang, Tianzhu
    Lu, Hanqing
    ADVANCES IN MULTIMEDIA MODELING, PROCEEDINGS, 2010, 5916 : 411 - 421
  • [5] Localized Multiple Kernel Learning for Realistic Human Action Recognition in Videos
    Song, Yan
    Zheng, Yan-Tao
    Tang, Sheng
    Zhou, Xiangdong
    Zhang, Yongdong
    Lin, Shouxun
    Chua, Tat-Seng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2011, 21 (09) : 1193 - 1202
  • [6] Human Action Recognition With Multiple-Instance Markov Model
    Zhou, Wen
    Zhang, Zhong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2014, 9 (10) : 1581 - 1591
  • [7] MULTIPLE INSTANCE DISCRIMINATIVE DICTIONARY LEARNING FOR ACTION RECOGNITION
    Li, Hongyang
    Chen, Jun
    Xu, Zengmin
    Chen, Huafeng
    Hu, Ruimin
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2014 - 2018
  • [8] Human Action Recognition Using Extreme Learning Machine via Multiple Types of Features
    Minhas, Rashid
    Baradarani, Aryaz
    Seifzadeh, Sepideh
    Wu, Q. M. Jonathan
    MOBILE MULTIMEDIA/IMAGE PROCESSING, SECURITY, AND APPLICATIONS 2010, 2010, 7708
  • [9] GRAPH-BASED MULTIPLE INSTANCE LEARNING FOR ACTION RECOGNITION
    Guo, Zixin
    Yi, Yang
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 3745 - 3749
  • [10] AN IMPROVED METHOD USING KINEMATIC FEATURES FOR ACTION RECOGNITION
    Chen, Yuanbo
    Zhao, Yanyun
    Cai, Anni
    PROCEEDINGS OF 2011 INTERNATIONAL CONFERENCE ON COMMUNICATION TECHNOLOGY AND APPLICATION, ICCTA2011, 2011, : 737 - 741