MMHAR-EnsemNet: A Multi-Modal Human Activity Recognition Model

被引:21
作者
Das, Avigyan [1 ]
Sil, Pritam [1 ]
Singh, Pawan Kumar [1 ]
Bhateja, Vikrant [2 ,3 ]
Sarkar, Ram [4 ]
机构
[1] Jadavpur Univ, Dept Informat Technol, Kolkata 700106, India
[2] Shri Ramswaroop Mem Grp Profess Coll SRMGPC, Dept Elect & Commun Engn, Lucknow 226028, Uttar Pradesh, India
[3] Dr APJ Abdul Kalam Tech Univ, Lucknow 226031, Uttar Pradesh, India
[4] Jadavpur Univ, Dept Comp Sci & Engn, Kolkata 700032, India
关键词
Skeleton; Data models; Activity recognition; Accelerometers; Gyroscopes; Three-dimensional displays; MMHAR-EnsemNet; human activity recognition; multi-modal data; skeleton; accelerometer; gyroscope; UTD-MHAD; Berkeley-MHAD;
D O I
10.1109/JSEN.2020.3034614
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this article, we propose a new deep learning model named as MMHAR-EnsemNet (Multi-Modal Human Activity Recognition Ensemble Network) which makes use of four different modalities to perform sensor-based Human Activity Recognition (HAR).Two separate Convolutional Neural Networks (CNNs) are made for skeleton data. While one CNN and one LSTM is trained for RGB images. For Accelerometer and Gyroscope data first it is converted to signal diagram then another CNN model is trained. Finally, all the outputs of the said models have been used to form an ensemble so that performance of the HAR model gets improved. The proposed model has been evaluated on two standard benchmark datasets namely UTD-MHAD and Berkeley-MHAD which contain four different modalities of input information. Experimental results confirm that the MMHAR-EnsemNet model has outperformed some recently proposed models considered here for comparison. Source code of this work can be found at: https://github.com/abhi1998das/MMHAREnsemNet.
引用
收藏
页码:11569 / 11576
页数:8
相关论文
共 34 条
  • [1] Human activity recognition from 3D data: A review
    Aggarwal, J. K.
    Xia, Lu
    [J]. PATTERN RECOGNITION LETTERS, 2014, 48 : 70 - 80
  • [2] Human Action Recognition Using Deep Multilevel Multimodal (M2) Fusion of Depth and Inertial Sensors
    Ahmad, Zeeshan
    Khan, Naimul
    [J]. IEEE SENSORS JOURNAL, 2020, 20 (03) : 1445 - 1455
  • [3] [Anonymous], 2013, Communications of the ACM, DOI [DOI 10.1145/2398356.2398381, 10.1145/2398356.2398381]
  • [4] Fuzzy Integral-Based CNN Classifier Fusion for 3D Skeleton Action Recognition
    Banerjee, Avinandan
    Singh, Pawan Kumar
    Sarkar, Ram
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (06) : 2206 - 2216
  • [5] Ben Mahjoub A, 2016, INT DES TEST SYMP, P83, DOI 10.1109/IDT.2016.7843019
  • [6] Multicomponent dark matter in extended U(1)B-L: neutrino mass and high scale validity
    Bhattacharya, Subhaditya
    Chakrabarty, Nabarun
    Roshan, Rishav
    Sil, Arunansu
    [J]. JOURNAL OF COSMOLOGY AND ASTROPARTICLE PHYSICS, 2020, (04):
  • [7] Chen C, 2015, IEEE IMAGE PROC, P168, DOI 10.1109/ICIP.2015.7350781
  • [8] Improving Human Action Recognition Using Fusion of Depth Camera and Inertial Sensors
    Chen, Chen
    Jafari, Roozbeh
    Kehtarnavaz, Nasser
    [J]. IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2015, 45 (01) : 51 - 61
  • [9] Novel Cadmium Responsive MicroRNAs in Daphnia pulex
    Chen, Shuai
    McKinney, Garrett J.
    Nichols, Krista M.
    Colbourne, John K.
    Sepulveda, Maria S.
    [J]. ENVIRONMENTAL SCIENCE & TECHNOLOGY, 2015, 49 (24) : 14605 - 14613
  • [10] A Novel Developmental Role for Dopaminergic Signaling to Specify Hypothalamic Neurotransmitter Identity
    Chen, Yu-Chia
    Semenova, Svetlana
    Rozov, Stanislav
    Sundvik, Maria
    Bonkowsky, Joshua L.
    Panula, Pertti
    [J]. JOURNAL OF BIOLOGICAL CHEMISTRY, 2016, 291 (42) : 21880 - 21892