Enhancing Human Action Recognition through Temporal Saliency

被引:0
|
作者
Adeli, Vida [1 ]
Fazl-Ersi, Ehsan [1 ]
Harati, Ahad [1 ]
机构
[1] Ferdowsi Univ Mashhad, Dept Comp Engn, Mashhad, Razavi Khorasan, Iran
来源
PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE (ICPRAI 2018) | 2018年
关键词
Action recognition; Motion; Region proposal; Convolutional Neural Networks; Actionness;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Images and videos have become ubiquitous in every aspects of life due to the growing digital recording devices. It has encouraged the development of algorithms that can analyze video content and perform human action recognition. This paper investigates the challenging problem of action recognition by outlining a new approach to represent a video sequence. A novel framework is developed to produce informative features for action labeling in a weakly-supervised learning (WSL) approach both during training and testing. Using appearance and motion information, the goal is to identify frame regions that are likely to contain actions. A three-stream convolutional neural network is adopted and improved by proposing a method based on extracting actionness regions. This results in less computation as it is processing only some parts of an RGB frame and also interpret less non-activity related regions, which can mislead the recognition system. We exploit UCF sports dataset as our evaluation benchmark, which is a dataset of realistic sports videos. We will show that our proposed approach could outperform other existing state-of-the art methods.
引用
收藏
页码:176 / 181
页数:6
相关论文
共 50 条
  • [21] RETHINKING TEMPORAL-RELATED SAMPLE FOR HUMAN ACTION RECOGNITION
    Wang, Jinpeng
    Li, Shiren
    Duan, Zhikui
    Yuan, Zhihao
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 2368 - 2372
  • [22] Spatio-temporal Video Autoencoder for Human Action Recognition
    Sousa e Santos, Anderson Carlos
    Pedrini, Helio
    PROCEEDINGS OF THE 14TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2019, : 114 - 123
  • [23] Spatiotemporal Saliency Representation Learning for Video Action Recognition
    Kong, Yongqiang
    Wang, Yunhong
    Li, Annan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1515 - 1528
  • [24] Action Recognition in Videos with Temporal Segments Fusions
    Fang, Yuanye
    Zhang, Rui
    Wang, Qiu-Feng
    Huang, Kaizhu
    ADVANCES IN BRAIN INSPIRED COGNITIVE SYSTEMS, 2020, 11691 : 244 - 253
  • [25] Enhancing human action recognition via structural average curves analysis
    Shichen Zeng
    Guoliang Lu
    Peng Yan
    Signal, Image and Video Processing, 2018, 12 : 1551 - 1558
  • [26] Enhancing human action recognition via structural average curves analysis
    Zeng, Shichen
    Lu, Guoliang
    Yan, Peng
    SIGNAL IMAGE AND VIDEO PROCESSING, 2018, 12 (08) : 1551 - 1558
  • [27] Spatial and temporal saliency based four-stream network with multi-task learning for action recognition
    Zong, Ming
    Wang, Ruili
    Ma, Yujun
    Ji, Wanting
    APPLIED SOFT COMPUTING, 2023, 132
  • [28] Human action recognition based on multi-mode spatial-temporal feature fusion
    Wang, Dongli
    Yang, Jun
    Zhou, Yan
    2019 22ND INTERNATIONAL CONFERENCE ON INFORMATION FUSION (FUSION 2019), 2019,
  • [29] Detecting action-relevant regions for action recognition using a three-stage saliency detection technique
    Wang, Xiaofang
    Qi, Chun
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (11-12) : 7413 - 7433
  • [30] Human Action Recognition Based on a Spatio-Temporal Video Autoencoder
    Sousa e Santos, Anderson Carlos
    Pedrini, Helio
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2020, 34 (11)