Detecting action tubes via spatial action estimation and temporal path inference

被引:4
|
作者
Li, Nannan [1 ]
Huang, Jingjia [1 ]
Li, Thomas [2 ]
Guo, Huiwen [3 ]
Li, Ge [1 ]
机构
[1] Peking Univ, Shenzhen Grad Sch, Sch Elect & Comp Engn, Beijing, Peoples R China
[2] Gpower Semicond Inc, Suzhou, Peoples R China
[3] Chinese Acad Sci, Shenzhen Inst Adv Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Action detection; Spatial localization; Region proposal network; Tracking-by-detection; SUM-PRODUCT NETWORKS; ACTION RECOGNITION;
D O I
10.1016/j.neucom.2018.05.033
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we address the problem of action detection in unconstrained video clips. Our approach starts from action detection on object proposals at each frame, then aggregates the frame-level detection results belonging to the same actor across the whole video via linking, associating, and tracking to generate action tubes that are spatially compact and temporally continuous. To achieve the target, a novel action detection model with two-stream architecture is firstly proposed, which utilizes the fused feature from both appearance and motion cues and can be trained end-to-end. Then, the association of the action paths is formulated as a maximum set coverage problem with the results of action detection as a priori. We utilize an incremental search algorithm to obtain all the action proposals at one-pass operation with great efficiency, especially while dealing with the video of long duration or with multiple action instances. Finally, a tracking-by-detection scheme is designed to further refine the generated action paths. Extensive experiments on three validation datasets, UCF-Sports, UCF-101 and J-HMDB, show that the proposed approach advances state-of-the-art action detection performance in terms of both accuracy and proposal quality. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:65 / 77
页数:13
相关论文
共 50 条
  • [31] SAST: Learning Semantic Action-Aware Spatial-Temporal Features for Efficient Action Recognition
    Wang, Fei
    Wang, Guorui
    Huang, Yunwen
    Chu, Hao
    IEEE ACCESS, 2019, 7 : 164876 - 164886
  • [32] Detecting Human Action as the Spatio-Temporal Tube of Maximum Mutual Information
    Wang, Taiqing
    Wang, Shengjin
    Ding, Xiaoqing
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2014, 24 (02) : 277 - 290
  • [33] EFFICIENT TEMPORAL-SPATIAL FEATURE GROUPING FOR VIDEO ACTION RECOGNITION
    Qiu, Zhikang
    Zhao, Xu
    Hu, Zhilan
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2176 - 2180
  • [34] Recurrent Spatial-Temporal Attention Network for Action Recognition in Videos
    Du, Wenbin
    Wang, Yali
    Qiao, Yu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (03) : 1347 - 1360
  • [35] Action Recognition Using a Spatial-Temporal Network for Wild Felines
    Feng, Liqi
    Zhao, Yaqin
    Sun, Yichao
    Zhao, Wenxuan
    Tang, Jiaxi
    ANIMALS, 2021, 11 (02): : 1 - 18
  • [36] A SPATIAL-TEMPORAL CONSTRAINT-BASED ACTION RECOGNITION METHOD
    Han, Tingting
    Yao, Hongxun
    Zhang, Yanhao
    Xu, Pengfei
    2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 2767 - 2771
  • [37] Spatial-Temporal Action Localization With Hierarchical Self-Attention
    Pramono, Rizard Renanda Adhi
    Chen, Yie-Tarng
    Fang, Wen-Hsien
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 625 - 639
  • [38] Using BlazePose on Spatial Temporal Graph Convolutional Networks for Action Recognition
    Alsawadi, Motasem S.
    El-Kenawy, El-Sayed M.
    Rio, Miguel
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 74 (01): : 19 - 36
  • [39] Multi-scale spatial–temporal convolutional neural network for skeleton-based action recognition
    Qin Cheng
    Jun Cheng
    Ziliang Ren
    Qieshi Zhang
    Jianming Liu
    Pattern Analysis and Applications, 2023, 26 (3) : 1303 - 1315
  • [40] Accelerating temporal action proposal generation via high performance computing
    Tian Wang
    Shiye Lei
    Youyou Jiang
    Choi Chang
    Hichem Snoussi
    Guangcun Shan
    Yao Fu
    Frontiers of Computer Science, 2022, 16