Detecting Human Action as the Spatio-Temporal Tube of Maximum Mutual Information

被引:18
|
作者
Wang, Taiqing [1 ,2 ]
Wang, Shengjin [1 ,2 ]
Ding, Xiaoqing [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing 100084, Peoples R China
[2] Tsinghua Natl Lab Informat Sci & Technol, Beijing 100084, Peoples R China
基金
国家高技术研究发展计划(863计划); 中国国家自然科学基金;
关键词
Action detection; feature trajectory; mutual information; spatio-temporal cuboid (ST-cuboid); spatio-temporal tube (ST-tube); RECOGNITION; MOTION; DENSE;
D O I
10.1109/TCSVT.2013.2276856
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Human action detection in complex scenes is a challenging problem due to its high-dimensional search space and dynamic backgrounds. To achieve efficient and accurate action detection, we represent a video sequence as a collection of feature trajectories and model human action as the spatio-temporal tube (ST-tube) of maximum mutual information. First, a random forest is built to evaluate the mutual information of feature trajectories toward the action class, and then a one-order Markov model is introduced to recursively infer the action regions at consecutive frames. By exploring the time-continuity property of feature trajectories, the action region is efficiently inferred at large temporal intervals. Finally, we obtain an ST-tube by concatenating the consecutive action regions bounding the human bodies. Compared with the popular spatio-temporal cuboid action model, the proposed ST-tube model is not only more efficient, but also more accurate in action localization. Experimental results on the KTH, CMU and UCF sports datasets validate the superiority of our approach over the state-of-the-art methods in both localization accuracy and time efficiency.
引用
收藏
页码:277 / 290
页数:14
相关论文
共 50 条
  • [21] Local and Global Context Reasoning for Spatio-Temporal Action Localization
    Ando, Ryuhei
    Babazaki, Yasunori
    Takahashi, Katsuhiko
    ADVANCES IN VISUAL COMPUTING, ISVC 2023, PT I, 2023, 14361 : 147 - 159
  • [22] Com-STAL: Compositional Spatio-Temporal Action Localization
    Wang, Shaomeng
    Yan, Rui
    Huang, Peng
    Dai, Guangzhao
    Song, Yan
    Shu, Xiangbo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7645 - 7657
  • [23] Detecting Hidden Objects Using Efficient Spatio-Temporal Knowledge Representation
    Olszewska, Joanna Isabelle
    AGENTS AND ARTIFICIAL INTELLIGENCE, ICAART 2016, 2017, 10162 : 302 - 313
  • [24] Improved Spatio-temporal Salient Feature Detection for Action Recognition
    Shabani, Amir H.
    Clausi, David A.
    Zelek, John S.
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2011, 2011,
  • [25] Dynamic Spatio-Temporal Bag of Expressions (D-STBoE) Model for Human Action Recognition
    Nazir, Saima
    Yousaf, Muhammad Haroon
    Nebel, Jean-Christophe
    Velastin, Sergio A.
    SENSORS, 2019, 19 (12)
  • [26] Relevance Detection in Cataract Surgery Videos by Spatio-Temporal Action Localization
    Ghamsarian, Negin
    Taschwer, Mario
    Putzgruber-Adamitsch, Doris
    Sarny, Stephanie
    Schoeffmann, Klaus
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 10720 - 10727
  • [27] Actor-Centric Spatio-Temporal Feature Extraction for Action Recognition
    Anil, Kunchala
    Bouroche, Melanie
    Schoen-Phelan, Bianca
    COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT I, 2024, 2009 : 586 - 599
  • [28] Online action proposal generation using spatio-temporal attention network
    Keisham, Kanchan
    Jalali, Amin
    Lee, Minho
    NEURAL NETWORKS, 2022, 153 : 518 - 529
  • [29] An End-to-End Spatio-Temporal Attention Model for Human Action Recognition from Skeleton Data
    Song, Sijie
    Lan, Cuiling
    Xing, Junliang
    Zeng, Wenjun
    Liu, Jiaying
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4263 - 4270
  • [30] Spatio-temporal dynamics of automatic processing of phonological information in visual words
    Wang, Xiao-Dong
    Wu, Yin-Yuan
    Liu, A. -Ping
    Wang, Peng
    SCIENTIFIC REPORTS, 2013, 3