Inertial Sensor Data to Image Encoding for Human Action Recognition

被引:35
作者
Ahmad, Zeeshan [1 ]
Khan, Naimul [1 ]
机构
[1] Ryerson Univ, Dept Elect & Comp Engn, Toronto, ON M5B 2K3, Canada
关键词
Support vector machines; Image recognition; Correlation; Inertial sensors; Computational modeling; Markov processes; Feature extraction; Deep learning; human action recognition; image encoding; mutimodal fusion; DEPTH; FUSION;
D O I
10.1109/JSEN.2021.3062261
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional Neural Networks (CNNs) are successful deep learning models in the field of computer vision. To get the maximum advantage of CNN model for Human Action Recognition (HAR) using inertial sensor data, in this paper, we use four types of spatial domain methods for transforming inertial sensor data to activity images, which are then utilized in a novel fusion framework. These four types of activity images are Signal Images (SI), Gramian Angular Field (GAF) Images, Markov Transition Field (MTF) Images and Recurrence Plot (RP) Images. Furthermore, for creating a multimodal fusion framework and to exploit activity images, we made each type of activity images multimodal by convolving with two spatial domain filters: Prewitt filter and High-boost filter. ResNet-18, a CNN model, is used to learn deep features from multi-modalities. Learned features are extracted from the last pooling layer of each ResNet and then fused by canonical correlation based fusion (CCF) for improving the accuracy of human action recognition. These highly informative features are served as input to a multi-class Support Vector Machine (SVM). Experimental results on three publicly available inertial datasets show the superiority of the proposed method over the current state-of-the-art.
引用
收藏
页码:10978 / 10988
页数:11
相关论文
共 53 条
[1]   Deep Learning for Heterogeneous Human Activity Recognition in Complex IoT Applications [J].
Abdel-Basset, Mohamed ;
Hawash, Hossam ;
Chang, Victor ;
Chakrabortty, Ripon K. ;
Ryan, Michael .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (08) :5653-5665
[2]   CNN-Based Multistage Gated Average Fusion (MGAF) for Human Action Recognition Using Depth and Inertial Sensors [J].
Ahmad, Zeeshan ;
Khan, Naimul .
IEEE SENSORS JOURNAL, 2021, 21 (03) :3623-3634
[3]   Human Action Recognition Using Convolutional Neural Network and Depth Sensor Data [J].
Ahmad, Zeeshan ;
Illanko, Kandasamy ;
Khan, Naimul ;
Androutsos, Dimitri .
2019 INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY AND COMPUTER COMMUNICATIONS (ITCC 2019), 2019, :1-5
[4]   Human Action Recognition Using Deep Multilevel Multimodal (M2) Fusion of Depth and Inertial Sensors [J].
Ahmad, Zeeshan ;
Khan, Naimul .
IEEE SENSORS JOURNAL, 2020, 20 (03) :1445-1455
[5]  
Ahmad Z, 2019, 2019 IEEE FIFTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2019), P429, DOI [10.1109/BigMM.2019.00026, 10.1109/BigMM.2019.00074]
[6]   Towards Improved Human Action Recognition Using Convolutional Neural Networks and Multimodal Fusion of Depth and Inertial Sensor Data [J].
Ahmad, Zeeshan ;
Khan, Naimul Mefraz .
2018 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM 2018), 2018, :223-230
[7]  
Alirezanejad M., 2014, INDIAN J SCI TECHNOL, V7, P517, DOI 10.17485/ijst/2014/v7i4.12
[8]  
Almaslukh B, 2017, INT J COMPUT SCI NET, V17, P160
[9]  
[Anonymous], 2011, P INT JOINT C ART IN, DOI DOI 10.5591/978-1-57735-516-8/IJCAI11-290
[10]   Activity recognition from user-annotated acceleration data [J].
Bao, L ;
Intille, SS .
PERVASIVE COMPUTING, PROCEEDINGS, 2004, 3001 :1-17