Human action recognition in immersive virtual reality based on multi-scale spatio-temporal attention network

被引:0
|
作者
Xiao, Zhiyong [1 ]
Chen, Yukun [1 ]
Zhou, Xinlei [1 ]
He, Mingwei [2 ]
Liu, Li [1 ]
Yu, Feng [1 ,2 ,3 ]
Jiang, Minghua [1 ,3 ]
机构
[1] Wuhan Text Univ, Sch Comp Sci & Artificial Intelligence, Wuhan, Peoples R China
[2] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[3] Engn Res Ctr Hubei Prov Clothing Informat, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
human activity recognition; multi-scale feature; spatio-temporal feature; virtual reality; SIMULATION; SENSORS;
D O I
暂无
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Wearable human action recognition (HAR) has practical applications in daily life. However, traditional HAR methods solely focus on identifying user movements, lacking interactivity and user engagement. This paper proposes a novel immersive HAR method called MovPosVR. Virtual reality (VR) technology is employed to create realistic scenes and enhance the user experience. To improve the accuracy of user action recognition in immersive HAR, a multi-scale spatio-temporal attention network (MSSTANet) is proposed. The network combines the convolutional residual squeeze and excitation (CRSE) module with the multi-branch convolution and long short-term memory (MCLSTM) module to extract spatio-temporal features and automatically select relevant features from action signals. Additionally, a multi-head attention with shared linear mechanism (MHASLM) module is designed to facilitate information interaction, further enhancing feature extraction and improving accuracy. The MSSTANet network achieves superior performance, with accuracy rates of 99.33% and 98.83% on the publicly available WISDM and PAMPA2 datasets, respectively, surpassing state-of-the-art networks. Our method showcases the potential to display user actions and position information in a virtual world, enriching user experiences and interactions across diverse application scenarios. image
引用
收藏
页数:15
相关论文
共 50 条
  • [1] A fast human action recognition network based on spatio-temporal features
    Xu, Jie
    Song, Rui
    Wei, Haoliang
    Guo, Jinhong
    Zhou, Yifei
    Huang, Xiwei
    NEUROCOMPUTING, 2021, 441 : 350 - 358
  • [2] Efficient spatio-temporal network for action recognition
    Su, Yanxiong
    Zhao, Qian
    JOURNAL OF REAL-TIME IMAGE PROCESSING, 2024, 21 (05)
  • [3] A Spatio-Temporal Multi-Scale Binary Descriptor
    Xompero, Alessio
    Lanz, Oswald
    Cavallaro, Andrea
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 4362 - 4375
  • [4] Try This for Size: Multi-Scale Teleportation in Immersive Virtual Reality
    Weissker, Tim
    Franzgrote, Matthis
    Kuhlen, Torsten
    IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2024, 30 (05) : 2298 - 2308
  • [5] Spatio-temporal multi-scale motion descriptor from a spatially-constrained decomposition for online action recognition
    Martinez, Fabio
    Manzanera, Antoine
    Romero, Eduardo
    IET COMPUTER VISION, 2017, 11 (07) : 541 - 549
  • [6] A Novel Spatio-Temporal Network of Multi-channel CNN and GCN for Human Activity Recognition Based on BAN
    Jianning Wu
    Qianghui Liu
    Neural Processing Letters, 2023, 55 : 11489 - 11507
  • [7] A Novel Spatio-Temporal Network of Multi-channel CNN and GCN for Human Activity Recognition Based on BAN
    Wu, Jianning
    Liu, Qianghui
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 11489 - 11507
  • [8] A Dual Pipeline With Spatio-Temporal Attention Fusion Approach for Human Activity Recognition
    Wang, Xiaodong
    Li, Ying
    Fang, Aiqing
    He, Pei
    Guo, Yangming
    IEEE SENSORS JOURNAL, 2024, 24 (15) : 25150 - 25162
  • [9] Multi-View Gait Recognition With Joint Local Multi-Scale and Global Contextual Spatio-Temporal Features
    Zhai, Wenzhe
    Li, Haomiao
    Zheng, Chaoqun
    Xing, Xianglei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1123 - 1135
  • [10] AROS: human action recognition by spatio-temporal fusion mechanism based on optimised subcarriers
    Tao, Zhiyong
    Guo, Xijun
    Liu, Ying
    INTERNATIONAL JOURNAL OF SENSOR NETWORKS, 2024, 45 (04) : 204 - 215