Cross-view action recognition understanding from exocentric to egocentric perspective

被引:0
|
作者
Truong, Thanh-Dat [1 ]
Luu, Khoa [1 ]
机构
[1] Univ Arkansas, Comp Vis & Image Understanding Lab, Fayetteville, AR 72701 USA
基金
美国国家科学基金会;
关键词
Cross-view action recognition; Self-attention; Egocentric action recognition; ATTENTION;
D O I
10.1016/j.neucom.2024.128731
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding action recognition in egocentric videos has emerged as a vital research topic with numerous practical applications. With the limitation in the scale of egocentric data collection, learning robust learning-based action recognition models remains difficult. Transferring knowledge learned from the scale exocentric data to the egocentric data is challenging due to the difference in videos across views. work introduces a novel cross-view learning approach to action recognition (CVAR) that effectively transfers knowledge from the exocentric to the selfish view. First, we present a novel geometric-based constraint the self-attention mechanism in Transformer based on analyzing the camera positions between two Then, we propose a new cross-view self-attention loss learned on unpaired cross-view data to enforce the attention mechanism learning to transfer knowledge across views. Finally, to further improve the performance of our cross-view learning approach, we present the metrics to measure the correlations in videos and attention maps effectively. Experimental results on standard egocentric action recognition benchmarks, i.e., Charades Ego, EPIC-Kitchens-55, and EPIC-Kitchens-100, have shown our approach's effectiveness and state-of-the-art performance.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Cross-View Generalisation in Action Recognition: Feature Design for Transitioning from Exocentric To Egocentric Views
    Rocha, Bernardo
    Moreno, Plinio
    Bernardino, Alexandre
    ROBOT 2023: SIXTH IBERIAN ROBOTICS CONFERENCE ADVANCES IN ROBOTICS, VOL 1, 2024, 976 : 155 - 166
  • [2] Cross-View Exocentric to Egocentric Video Synthesis
    Liu, Gaowen
    Tang, Hao
    Latapie, Hugo M.
    Corso, Jason J.
    Yan, Yan
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 974 - 982
  • [3] Cross-view Action Modeling, Learning and Recognition
    Wang, Jiang
    Nie, Xiaohan
    Xia, Yin
    Wu, Ying
    Zhu, Song-Chun
    2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 2649 - 2656
  • [4] MODELING CROSS-VIEW INTERACTION CONSISTENCY FOR PAIRED EGOCENTRIC INTERACTION RECOGNITION
    Li, Zhongguo
    Lyu, Fan
    Feng, Wei
    Wang, Song
    2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,
  • [5] Cross-View Action Recognition via View Knowledge Transfer
    Liu, Jingen
    Shah, Mubarak
    Kuipers, Benjamin
    Savarese, Silvio
    2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011,
  • [6] Cross-view action recognition by cross-domain learning
    Nie, Weizhi
    Liu, Anan
    Li, Wenhui
    Su, Yuting
    IMAGE AND VISION COMPUTING, 2016, 55 : 109 - 118
  • [7] Cross-View Action Recognition from Temporal Self-similarities
    Junejo, Imran N.
    Dexter, Emilie
    Laptev, Ivan
    Perez, Patrick
    COMPUTER VISION - ECCV 2008, PT II, PROCEEDINGS, 2008, 5303 : 293 - 306
  • [8] Discriminative Virtual Views for Cross-View Action Recognition
    Li, Ruonan
    Zickler, Todd
    2012 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2012, : 2855 - 2862
  • [9] Heterogeneous discriminant analysis for cross-view action recognition
    Sui, Wanchen
    Wu, Xinxiao
    Feng, Yang
    Jia, Yunde
    NEUROCOMPUTING, 2016, 191 : 286 - 295
  • [10] Adversarial Feature Refinement for Cross-View Action Recognition
    Marsella, Antonio
    Goyal, Gaurvi
    Odone, Francesca
    36TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2021, 2021, : 1046 - 1054