Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning

被引:277
|
作者
Ali, Saad [1 ]
Shah, Mubarak [2 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[2] Univ Cent Florida, Comp Vis Lab, Sch Elect Engn & Comp Sci, Harris Corp Engn Ctr, Orlando, FL 32816 USA
关键词
Action recognition; motion; video analysis; principal component analysis; kinematic features;
D O I
10.1109/TPAMI.2008.284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a set of kinematic features that are derived from the optical flow for human action recognition in videos. The set of kinematic features includes divergence, vorticity, symmetric and antisymmetric flow fields, second and third principal invariants of flow gradient and rate of strain tensor, and third principal invariant of rate of rotation tensor. Each kinematic feature, when computed from the optical flow of a sequence of images, gives rise to a spatiotemporal pattern. It is then assumed that the representative dynamics of the optical flow are captured by these spatiotemporal patterns in the form of dominant kinematic trends or kinematic modes. These kinematic modes are computed by performing Principal Component Analysis (PCA) on the spatiotemporal volumes of the kinematic features. For classification, we propose the use of multiple instance learning (MIL) in which each action video is represented by a bag of kinematic modes. Each video is then embedded into a kinematic-mode-based feature space and the coordinates of the video in that space are used for classification using the nearest neighbor algorithm. The qualitative and quantitative results are reported on the benchmark data sets.
引用
收藏
页码:288 / 303
页数:16
相关论文
共 50 条
  • [41] Robust Human Action Recognition Using Dynamic Movement Features
    Zhang, Huiwen
    Fu, Mingliang
    Luo, Haitao
    Zhou, Weijia
    INTELLIGENT ROBOTICS AND APPLICATIONS, ICIRA 2017, PT I, 2017, 10462 : 474 - 484
  • [42] LEARNING DEEP TRAJECTORY DESCRIPTOR FOR ACTION RECOGNITION IN VIDEOS USING DEEP NEURAL NETWORKS
    Shi, Yemin
    Zeng, Wei
    Huang, Tiejun
    Wang, Yaowei
    2015 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO (ICME), 2015,
  • [43] Efficient large-scale action recognition in videos using extreme learning machines
    Varol, Gul
    Salah, Albert Ali
    EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (21) : 8274 - 8282
  • [44] Action density based frame sampling for human action recognition in videos
    Lin, Jie
    Mu, Zekun
    Zhao, Tianqing
    Zhang, Hanlin
    Yang, Xinyu
    Zhao, Peng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 90
  • [45] HUMAN ACTION RECOGNITION USING ROBUST POWER SPECTRUM FEATURES
    Ragheb, Hossein
    Velastin, Sergio
    Remagnino, Paolo
    Ellis, Tim
    2008 15TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1-5, 2008, : 753 - 756
  • [46] EFFICIENT POOLING OF IMAGE BASED CNN FEATURES FOR ACTION RECOGNITION IN VIDEOS
    Banerjee, Biplab
    Murino, Vittorio
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2637 - 2641
  • [47] Human Action Recognition from RGBD Videos based on Retina Model and Local Binary Pattern Features
    Al-Akam, Rawya
    Al-Darraji, Salah
    Paulus, Dietrich
    26. INTERNATIONAL CONFERENCE IN CENTRAL EUROPE ON COMPUTER GRAPHICS, VISUALIZATION AND COMPUTER VISION (WSCG 2018), 2018, 2803 : 1 - 7
  • [48] Action Recognition Based on Linear Dynamical Systems with Deep Features in Videos
    Du, Zhouning
    Mukaidani, Hiroaki
    Saravanakumar, Ramasamy
    2020 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2020, : 2634 - 2639
  • [49] Human Action Recognition in First Person Videos using Verb-Object Pairs
    Gokce, Zeynep
    Pehlivan, Selen
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [50] Human Action Recognition in Unconstrained Videos by Explicit Motion Modeling
    Jiang, Yu-Gang
    Dai, Qi
    Liu, Wei
    Xue, Xiangyang
    Ngo, Chong-Wah
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (11) : 3781 - 3795