Cross-Modality Self-Attention and Fusion-Based Neural Network for Lower Limb Locomotion Mode Recognition

被引:2
作者
Zhao, Changchen [1 ]
Liu, Kai [2 ]
Zheng, Hao [3 ]
Song, Wenbo [4 ]
Pei, Zhongcai [3 ]
Chen, Weihai [5 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci, Hangzhou 310018, Peoples R China
[2] Zhejiang Univ Technol, Coll Informat Engn, Hangzhou 310023, Peoples R China
[3] Beihang Univ, Hangzhou Innovat Inst, Hangzhou 310051, Peoples R China
[4] Jilin Normal Univ, Coll Phys Educ, Siping 136000, Peoples R China
[5] Anhui Univ, Sch Elect Engn & Automat, Hefei 230601, Peoples R China
关键词
Cross-modality interaction; self-attention; locomotion mode recognition; lower limb; neural network; INTENT RECOGNITION; PREDICTION; STRATEGY; GAZE;
D O I
10.1109/TASE.2024.3421276
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Although there are many wearable sensors that make the acquisition of multi-modality data easier, effective feature extraction and fusion of the data is still challenging for lower limb locomotion mode recognition. In this article, a novel neural network is proposed for accurate prediction of five common lower limb locomotion modes including level walking, ramp ascent, ramp descent, stair ascent, and stair descent. First, the encoder-decoder structure is employed to enrich the channel diversity for the separation of the useful patterns from combined patterns. Second, a self-attention based cross-modality interaction module is proposed, which enables bilateral information flow between two encoding paths to fully exploit the interdependencies and to find complementary information between modalities. Third, a multi-modality fusion module is designed where the complementary features are fused by a channel-wise weighted summation whose coefficients are learned end-to-end. A benchmark dataset is collected from 10 health subjects containing EMG and IMU signals and five locomotion modes. Extensive experiments are conducted on one publicly available dataset ENABL3S and one self-collected dataset. The results show that the proposed method outperforms the compared methods with higher classification accuracy. The proposed method achieves a classification accuracy of 98.25 $\%$ on ENABL3S dataset and 95.51 $\%$ on the self-collected dataset. Note to Practitioners-This article aims to solve the real challenges encountered when intelligent recognition algorithms are applied in wearable robots: how to effectively and efficiently fuse the multi-modality data for better decision-making. First, most existing methods directly concatenate the multi-modality data, which increases the data dimensionality and brings computational burden. Second, existing recognition neural networks continuously compress the feature size such that the discriminative patterns are submerged in the noise and thus difficult to be identified. This research decomposes the mixed input signals on the channel dimension such that the useful patterns can be separated. Moreover, this research employs self-attention mechanism to associate correlations between two modalities and use this correlation as a new feature for subsequent representation learning, generating new, compact, and complementary features for classification. We demonstrate that the proposed network achieves 98.25 $\%$ accuracy and 3.5 ms prediction time. We anticipate that the proposed network could be a general scientific and practical methodology of multi-modality signal fusion and feature learning for intelligent systems.
引用
收藏
页码:5411 / 5424
页数:14
相关论文
共 50 条
  • [21] SelfGCN: Graph Convolution Network With Self-Attention for Skeleton-Based Action Recognition
    Wu, Zhize
    Sun, Pengpeng
    Chen, Xin
    Tang, Keke
    Xu, Tong
    Zou, Le
    Wang, Xiaofeng
    Tan, Ming
    Cheng, Fan
    Weise, Thomas
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4391 - 4403
  • [22] Self-attention convolutional neural network based fault diagnosis algorithm for chemical process
    Ren Jia
    Zou Hongrui
    Tang Lijuan
    Sun Siyu
    Shen Qihao
    Wang Xiang
    Bao Ke
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 4046 - 4051
  • [23] A time-aware self-attention based neural network model for sequential recommendation
    Zhang, Yihu
    Yang, Bo
    Liu, Haodong
    Li, Dongsheng
    APPLIED SOFT COMPUTING, 2023, 133
  • [24] A strain gauge based locomotion mode recognition method using convolutional neural network
    Feng, Yanggang
    Chen, Wanwen
    Wang, Qining
    ADVANCED ROBOTICS, 2019, 33 (05) : 254 - 263
  • [25] Multiscale Temporal Self-Attention and Dynamical Graph Convolution Hybrid Network for EEG-Based Stereogram Recognition
    Shen, Lili
    Sun, Mingyang
    Li, Qunxia
    Li, Beichen
    Pan, Zhaoqing
    Lei, Jianjun
    IEEE TRANSACTIONS ON NEURAL SYSTEMS AND REHABILITATION ENGINEERING, 2022, 30 : 1191 - 1202
  • [26] A class alignment network based on self-attention for cross-subject EEG classification
    Ma, Sufan
    Zhang, Dongxiao
    Wang, Jiayi
    Xie, Jialiang
    BIOMEDICAL PHYSICS & ENGINEERING EXPRESS, 2025, 11 (01):
  • [27] A Novel Cross-Attention Fusion-Based Joint Training Framework for Robust Underwater Acoustic Signal Recognition
    Zhou, Aolong
    Li, Xiaoyong
    Zhang, Wen
    Li, Dawei
    Deng, Kefeng
    Ren, Kaijun
    Song, Junqiang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [28] Integral Real-time Locomotion Mode Recognition Based on GA-CNN for Lower Limb Exoskeleton
    Wang, Jiaqi
    Wu, Dongmei
    Gao, Yongzhuo
    Wang, Xinrui
    Li, Xiaoqi
    Xu, Guoqiang
    Dong, Wei
    JOURNAL OF BIONIC ENGINEERING, 2022, 19 (05) : 1359 - 1373
  • [29] Integral Real-time Locomotion Mode Recognition Based on GA-CNN for Lower Limb Exoskeleton
    Jiaqi Wang
    Dongmei Wu
    Yongzhuo Gao
    Xinrui Wang
    Xiaoqi Li
    Guoqiang Xu
    Wei Dong
    Journal of Bionic Engineering, 2022, 19 : 1359 - 1373
  • [30] SACall: A Neural Network Basecaller for Oxford Nanopore Sequencing Data Based on Self-Attention Mechanism
    Huang, Neng
    Nie, Fan
    Ni, Peng
    Luo, Feng
    Wang, Jianxin
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2022, 19 (01) : 614 - 623