Egocentric action anticipation from untrimmed videos

被引:0
|
作者
Rodin, Ivan [1 ]
Furnari, Antonino [1 ,2 ]
Farinella, Giovanni Maria [1 ,2 ]
机构
[1] Univ Catania, Catania, Italy
[2] Univ Catania, Next Vis srl Spinoff, Catania, Italy
关键词
computer vision; pattern recognition;
D O I
10.1049/cvi2.12342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Egocentric action anticipation involves predicting future actions performed by the camera wearer from egocentric video. Although the task has recently gained attention in the research community, current approaches often assume that input videos are 'trimmed', meaning that a short video sequence is sampled a fixed time before the beginning of the action. However, trimmed action anticipation has limited applicability in real-world scenarios, where it is crucial to deal with 'untrimmed' video inputs and the exact moment of action initiation cannot be assumed at test time. To address these limitations, an untrimmed action anticipation task is proposed, which, akin to temporal action detection, assumes that the input video is untrimmed at test time, while still requiring predictions to be made before actions take place. The authors introduce a benchmark evaluation procedure for methods designed to address this novel task and compare several baselines on the EPIC-KITCHENS-100 dataset. Through our experimental evaluation, testing a variety of models, the authors aim to better understand their performance in untrimmed action anticipation. Our results reveal that the performance of current models designed for trimmed action anticipation is limited, emphasising the need for further research in this area.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Rethinking Online Action Detection in Untrimmed Videos: A Novel Online Evaluation Protocol
    Baptista-Rios, Marcos
    Lopez-Sastre, Roberto J.
    Caba Heilbron, Fabian
    Van Gemert, Jan C.
    Acevedo-Rodriguez, F. Javier
    Maldonado-Bascon, Saturnino
    IEEE ACCESS, 2020, 8 : 5139 - 5146
  • [2] Holographic Feature Learning of Egocentric-Exocentric Videos for Multi-Domain Action Recognition
    Huang, Yi
    Yang, Xiaoshan
    Gao, Junyun
    Xu, Changsheng
    IEEE Transactions on Multimedia, 2022, 24 : 2273 - 2286
  • [3] An Effective and Efficient Method for Detecting Hands in Egocentric Videos for Rehabilitation Applications
    Visee, Ryan J.
    Likitlersuang, Jirapat
    Zariffa, Jose
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2020, 28 (03) : 748 - 755
  • [4] Egocentric Vision-based Action Recognition: A survey
    Nunez-Marcos, Adrian
    Azkune, Gorka
    Arganda-Carreras, Ignacio
    NEUROCOMPUTING, 2022, 472 : 175 - 197
  • [5] Structured Learning for Action Recognition in Videos
    Long, Yinghan
    Srinivasan, Gopalakrishnan
    Panda, Priyadarshini
    Roy, Kaushik
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (03) : 475 - 484
  • [6] ASTRA: An Action Spotting TRAnsformer for Soccer Videos
    Xarles, Artur
    Escalera, Sergio
    Moeslund, Thomas B.
    Clapes, Albert
    PROCEEDINGS OF THE 6TH INTERNATIONAL WORKSHOP ON MULTIMEDIA CONTENT ANALYSIS IN SPORTS, MMSPORTS 2023, 2023, : 93 - 102
  • [7] Autoencoder-based unsupervised one-class learning for abnormal activity detection in egocentric videos
    Hu, Haowen
    Hachiuma, Ryo
    Saito, Hideo
    IET COMPUTER VISION, 2025, 19 (01)
  • [8] A Holistic Approach for Role Inference and Action Anticipation in Human Teams
    Dong, Junyi
    Huo, Qingze
    Ferrari, Silvia
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2022, 13 (06)
  • [9] LONG-TERM ACTION ANTICIPATION BASED ON CONTEXTUAL ALIGNMENT
    Patsch, Constantin
    Zhang, Jinghan
    Wu, Yuankai
    Zakour, Marsil
    Salihu, Driton
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 5920 - 5924
  • [10] Automatic excavator action recognition and localisation for untrimmed video using hybrid LSTM-Transformer networks
    Martin, Abbey
    Hill, Andrew J.
    Seiler, Konstantin M.
    Balamurali, Mehala
    INTERNATIONAL JOURNAL OF MINING RECLAMATION AND ENVIRONMENT, 2024, 38 (05) : 353 - 372