Multihead-Res-SE Residual Network with Attention for Human Activity Recognition

被引:0
|
作者
Kang, Hongbo [1 ]
Lv, Tailong [1 ]
Yang, Chunjie [1 ]
Wang, Wenqing [1 ]
机构
[1] Xian Univ Posts & Telecommun, Sch Automat, Xian 710100, Peoples R China
关键词
human activity recognition; deep learning; residual block; squeeze-and-excitation module; multichannel CNN; attention mechanism; WEARABLE SENSOR;
D O I
10.3390/electronics13173407
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human activity recognition (HAR) typically uses wearable sensors to identify and analyze the time-series data they collect, enabling recognition of specific actions. As such, HAR is increasingly applied in human-computer interaction, healthcare, and other fields, making accurate and efficient recognition of various human activities. In recent years, deep learning methods have been extensively applied in sensor-based HAR, yielding remarkable results. However, complex HAR research, which involves specific human behaviors in varied contexts, still faces several challenges. To solve these problems, we propose a multi-head neural network based on the attention mechanism. This framework contains three convolutional heads, with each head designed using one-dimensional CNN to extract features from sensory data. The model uses a channel attention module (squeeze-excitation module) to enhance the representational capabilities of convolutional neural networks. We conducted experiments on two publicly available benchmark datasets, UCI-HAR and WISDM, to evaluate our model. The results were satisfactory, with overall recognition accuracies of 96.72% and 97.73% on their respective datasets. The experimental results demonstrate the effectiveness of the network structure for the HAR, which ensures a higher level of accuracy.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Convolutional Neural Network With Multihead Attention for Human Activity Recognition
    Tan, Tan-Hsu
    Chang, Yang-Lang
    Wu, Jun-Rong
    Chen, Yung-Fu
    Alkhaleefah, Mohammad
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (02): : 3032 - 3043
  • [2] Attention-Based Residual BiLSTM Networks for Human Activity Recognition
    Zhang, Junjie
    Liu, Yuanhao
    Yuan, Hua
    IEEE ACCESS, 2023, 11 : 94173 - 94187
  • [3] Attention induced multi-head convolutional neural network for human activity recognition
    Khan, Zanobya N.
    Ahmad, Jamil
    APPLIED SOFT COMPUTING, 2021, 110
  • [4] A Novel IoT-Perceptive Human Activity Recognition (HAR) Approach Using Multihead Convolutional Attention
    Zhang, Haoxi
    Xiao, Zhiwen
    Wang, Juan
    Li, Fei
    Szczerbicki, Edward
    IEEE INTERNET OF THINGS JOURNAL, 2020, 7 (02): : 1072 - 1080
  • [5] Multi-ResAtt: Multilevel Residual Network With Attention for Human Activity Recognition Using Wearable Sensors
    Al-qaness, Mohammed A. A.
    Dahou, Abdelghani
    Abd Elaziz, Mohamed
    Helmi, A. M.
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2023, 19 (01) : 144 - 152
  • [6] A Novel Attention-Based Convolution Neural Network for Human Activity Recognition
    Zheng, Ge
    IEEE SENSORS JOURNAL, 2021, 21 (23) : 27015 - 27025
  • [7] Sensors-based Human Activity Recognition with Convolutional Neural Network and Attention Mechanism
    Zhang, Wenbo
    Zhu, Tao
    Yang, Congmin
    Xiao, Jiyi
    Ning, Huansheng
    PROCEEDINGS OF 2020 IEEE 11TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND SERVICE SCIENCE (ICSESS 2020), 2020, : 158 - 162
  • [8] Human Activity Recognition Using WiFi Signal Features and Efficient Residual Packet Attention Network
    Yang, Senquan
    Yang, Junjie
    Yang, Chao
    Yan, Wei
    Li, Pu
    IEEE SENSORS LETTERS, 2025, 9 (04)
  • [9] A Shallow Convolution Network Based Contextual Attention for Human Activity Recognition
    Xu, Chenyang
    Mao, Zhihong
    Fan, Feiyi
    Qiu, Tian
    Shen, Jianfei
    Gu, Yang
    MOBILE AND UBIQUITOUS SYSTEMS: COMPUTING, NETWORKING AND SERVICES, MOBIQUITOUS 2022, 2023, 492 : 155 - 171
  • [10] Attention-based LSTM Network for Wearable Human Activity Recognition
    Sun, Bo
    Liu, Meiqin
    Zheng, Ronghao
    Zhang, Senlin
    PROCEEDINGS OF THE 38TH CHINESE CONTROL CONFERENCE (CCC), 2019, : 8677 - 8682