Detecting Human Action as the Spatio-Temporal Tube of Maximum Mutual Information

被引:18
|
作者
Wang, Taiqing [1 ,2 ]
Wang, Shengjin [1 ,2 ]
Ding, Xiaoqing [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[2] Tsinghua Natl Lab Informat Sci & Technol, Beijing 100084, Peoples R China
基金
国家高技术研究发展计划(863计划); 中国国家自然科学基金;
关键词
Action detection; feature trajectory; mutual information; spatio-temporal cuboid (ST-cuboid); spatio-temporal tube (ST-tube); RECOGNITION; MOTION; DENSE;
D O I
10.1109/TCSVT.2013.2276856
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Human action detection in complex scenes is a challenging problem due to its high-dimensional search space and dynamic backgrounds. To achieve efficient and accurate action detection, we represent a video sequence as a collection of feature trajectories and model human action as the spatio-temporal tube (ST-tube) of maximum mutual information. First, a random forest is built to evaluate the mutual information of feature trajectories toward the action class, and then a one-order Markov model is introduced to recursively infer the action regions at consecutive frames. By exploring the time-continuity property of feature trajectories, the action region is efficiently inferred at large temporal intervals. Finally, we obtain an ST-tube by concatenating the consecutive action regions bounding the human bodies. Compared with the popular spatio-temporal cuboid action model, the proposed ST-tube model is not only more efficient, but also more accurate in action localization. Experimental results on the KTH, CMU and UCF sports datasets validate the superiority of our approach over the state-of-the-art methods in both localization accuracy and time efficiency.
引用
收藏
页码:277 / 290
页数:14
相关论文
共 50 条
  • [41] A Moving Target Detection Model Inspired by Spatio-Temporal Information Accumulation of Avian Tectal Neurons
    Huang, Shuman
    Niu, Xiaoke
    Wang, Zhizhong
    Liu, Gang
    Shi, Li
    MATHEMATICS, 2023, 11 (05)
  • [42] Deep Spatio-temporal Beam Training for mmWave Communications with Human Self-blockage
    Shan, Wenxing
    Ma, Yiming
    Wang, Zicun
    Zhang, Lin
    Xiao, Ming
    2023 IEEE 98TH VEHICULAR TECHNOLOGY CONFERENCE, VTC2023-FALL, 2023,
  • [43] Spatio-Temporal Frames in a Bag-of-visual-features Approach for Human Actions Recognition
    Lopes, Ana Paula B.
    Oliveira, Rodrigo S.
    de Almeida, Jussara M.
    Araujo, Arnaldo de A.
    2009 XXII BRAZILIAN SYMPOSIUM ON COMPUTER GRAPHICS AND IMAGE PROCESSING (SIBGRAPI 2009), 2009, : 315 - 321
  • [44] Fluxformer: Flow-Guided Duplex Attention Transformer via Spatio-Temporal Clustering for Action Recognition
    Hong, Younggi
    Kim, Min Ju
    Lee, Isack
    Yoo, Seok Bong
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (10) : 6411 - 6418
  • [45] ADVANCES ON ACTION RECOGNITION IN VIDEOS USING AN INTEREST POINT DETECTOR BASED ON MULTIBAND SPATIO-TEMPORAL ENERGIES
    Maninis, Kevis
    Koutras, Petros
    Maragos, Petros
    2014 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2014, : 1490 - 1494
  • [46] SPATIO-TEMPORAL REPRESENTATION FOR FACE AUTHENTICATION BY USING MULTI-TASK LEARNING WITH HUMAN ATTRIBUTES
    Kim, Seong Tae
    Kim, Dae Hoe
    Ro, Yong Man
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 2996 - 3000
  • [47] A KPCA spatio-temporal differential geometric trajectory cloud classifier for recognizing human actions in a CBVR system
    Gomez-Conde, Ivan
    Olivieri, David N.
    EXPERT SYSTEMS WITH APPLICATIONS, 2015, 42 (13) : 5472 - 5490
  • [48] Human behaviour analysis based on spatio-temporal dual-stream heterogeneous convolutional neural network
    Ye, Qing
    Zhao, Yuqi
    Zhong, Haoxin
    INTERNATIONAL JOURNAL OF COMPUTATIONAL SCIENCE AND ENGINEERING, 2023, 26 (06) : 673 - 683
  • [49] Spatio-Temporal Representation Matching-Based Open-Set Action Recognition by Joint Learning of Motion and Appearance
    Yoon, Yongsang
    Yu, Jongmin
    Jeon, Moongu
    IEEE ACCESS, 2019, 7 : 165997 - 166010
  • [50] Spatio-Temporal Convolutional Networks and N-Ary Ontologies for Human Activity-Aware Robotic System
    Abdelkawy, H.
    Ayari, N.
    Chibani, A.
    Amirat, Y.
    Attal, F.
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02): : 620 - 627