Gesture Recognition Using MLP-Mixer With CNN and Stacking Ensemble for sEMG Signals

被引:6
作者
Shen, Shu [1 ,2 ]
Li, Minglei [1 ]
Mao, Fan [1 ]
Chen, Xinrong [3 ]
Ran, Ran [1 ]
机构
[1] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing 210023, Jiangsu, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Jiangsu High Technol Res Key Lab Wireless Sensor, Nanjing 210023, Jiangsu, Peoples R China
[3] Fudan Univ, Acad Engn & Technol, Shanghai 200032, Peoples R China
基金
中国国家自然科学基金;
关键词
Convolution; Gesture recognition; Feature extraction; Kernel; Convolutional neural networks; Stacking; Sensors; Convolutional neural network (CNN); deep learning; ensemble learning; gesture recognition; human-computer interaction (HCI); multilayer perceptron (MLP)-Mixer; surface electromyography (sEMG);
D O I
10.1109/JSEN.2023.3347529
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In recent years, gesture perception has become crucial to human-computer interaction (HCI) technologies. Among various techniques, gesture recognition based on surface electromyography (sEMG) signals has gained significant prominence, with deep-learning methods playing a pivotal role in this domain. However, as the demand for accurate gesture recognition continues to rise, there is a growing inclination toward selecting complex deep neural network architectures. This trend, however, poses challenges in terms of performance and runtime requirements for computing devices. This article introduces a novel gesture recognition method utilizing the multilayer perceptron (MLP)-Mixer framework combined with Stacking ensemble learning to address these challenges. The proposed method effectively captures the features of sEMG data by employing simple MLPs, achieving a level of accuracy comparable to complex networks while simultaneously reducing inference time. Experimental results demonstrate that the method performs a classification accuracy of 80.03% and 81.13% for 49 actions in the open-source dataset NinaPro DB2, using window lengths of 200 and 300 ms, respectively. Furthermore, the method achieves a single inference speed of 54.77 ms with a window length of 200 ms. In NinaPro DB5, with window lengths of 250 and 300 ms, the method presented in this article achieves accuracy rates of 73.39% and 74.82%, respectively, completing inference in just 11.45 ms using the 300-ms window length. Notably, the technique also demonstrates its ability to mitigate the impact of individual differences in sEMG data on recognition accuracy.
引用
收藏
页码:4960 / 4968
页数:9
相关论文
共 33 条
  • [1] Abdelouahad A., 2018, P 4 INT C OPT APPL I, P1
  • [2] Deep Learning with Convolutional Neural Networks Applied to Electromyography Data: A Resource for the Classification of Movements for Prosthetic Hands
    Atzori, Manfredo
    Cognolato, Matteo
    Mueller, Henning
    [J]. FRONTIERS IN NEUROROBOTICS, 2016, 10
  • [3] Electromyography data for non-invasive naturally-controlled robotic hand prostheses
    Atzori, Manfredo
    Gijsberts, Arjan
    Castellini, Claudio
    Caputo, Barbara
    Hager, Anne-Gabrielle Mittaz
    Elsig, Simone
    Giatsidis, Giorgio
    Bassetto, Franco
    Muller, Henning
    [J]. SCIENTIFIC DATA, 2014, 1
  • [4] Assessing fractal dimension methods as feature extractors for EMG signal classification
    Coelho, Andre L. V.
    Lima, Clodoaldo A. M.
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2014, 36 : 81 - 98
  • [5] Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
  • [6] Fattah S. A., 2012, INT J SIGNAL IMAGE P, V3, P99, DOI [DOI 10.5121/SIPIJ.2012.3207, 10.5121/sipij.2012.3207]
  • [7] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [8] He YN, 2018, IEEE ENG MED BIO, P5636, DOI 10.1109/EMBC.2018.8513595
  • [9] Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
  • [10] A novel attention-based hybrid CNN-RN N architecture for sEMG-based gesture recognition
    Hu, Yu
    Wong, Yongkang
    Wei, Wentao
    Du, Yu
    Kankanhalli, Mohan
    Geng, Weidong
    [J]. PLOS ONE, 2018, 13 (10):