C2FHAR: Coarse-to-Fine Human Activity Recognition With Behavioral Context Modeling Using Smart Inertial Sensors

被引:17
作者
Ehatisham-Ul-Haq, Muhammad [1 ]
Azam, Muhammad Awais [1 ,2 ]
Amin, Yasar [1 ]
Naeem, Usman [3 ]
机构
[1] UET, Fac Telecom & Informat Engn, Taxila 47050, Pakistan
[2] Whitecliffe Technol, Fac Informat Technol, Wellington 6011, New Zealand
[3] Queen Mary Univ London, Sch Elect Engn & Comp Sci, Fac Sci & Engn, London E1 4NS, England
来源
IEEE ACCESS | 2020年 / 8卷
关键词
Activity recognition; behavioral context; context-aware; machine learning; smart sensing; ACCELEROMETER DATA; WEARABLE SENSORS; DATA FUSION; MOBILE; CLASSIFICATION; ALGORITHMS; NETWORKS; FEATURES; SYSTEM;
D O I
10.1109/ACCESS.2020.2964237
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Smart sensing devices are furnished with an array of sensors, including locomotion sensors, which enable continuous and passive monitoring of human activities for the ambient assisted living. As a result, sensor-based human activity recognition has earned significant popularity in the past few years. A lot of successful research studies have been conducted in this regard. However, the accurate recognition of <italic>in-the-wild</italic> human activities in real-time is still a fundamental challenge to be addressed as human physical activity patterns are adversely affected by their behavioral contexts. Moreover, it is essential to infer a user & x2019;s behavioral context along with the physical activity to enable context-aware and knowledge-driven applications in real-time. Therefore, this research work presents & x201C;C2FHAR & x201D;, a novel approach for <italic>coarse-to-fine human activity recognition in-the-wild</italic>, which explicitly models the user & x2019;s behavioral contexts with activities of daily living to learn and recognize the fine-grained human activities. For addressing real-time activity recognition challenges, the proposed scheme utilizes a multi-label classification model for identifying <italic>in-the-wild</italic> human activities at two different levels, i.e., <italic>coarse</italic> or <italic>fine-grained</italic>, depending upon the real-time use-cases. The proposed scheme is validated with extensive experiments using heterogeneous sensors, which demonstrate its efficacy.
引用
收藏
页码:7731 / 7747
页数:17
相关论文
共 70 条
  • [1] Activity Recognition with Evolving Data Streams: A Review
    Abdallah, Zahraa S.
    Gaber, Mohamed Medhat
    Srinivasan, Bala
    Krishnaswamy, Shonali
    [J]. ACM COMPUTING SURVEYS, 2018, 51 (04)
  • [2] [Anonymous], 2016, P INT JOINT C ART IN
  • [3] [Anonymous], J AMBIENT INTELL HUM
  • [4] [Anonymous], C4 5 PROGRAMS MACHIN
  • [5] Hand, belt, pocket or bag: Practical activity tracking with mobile phones
    Antos, Stephen A.
    Albert, Mark V.
    Kording, Konrad P.
    [J]. JOURNAL OF NEUROSCIENCE METHODS, 2014, 231 : 22 - 30
  • [6] Physical Human Activity Recognition Using Wearable Sensors
    Attal, Ferhat
    Mohammed, Samer
    Dedabrishvili, Mariam
    Chamroukhi, Faicel
    Oukhellou, Latifa
    Amirat, Yacine
    [J]. SENSORS, 2015, 15 (12) : 31314 - 31338
  • [7] Activity recognition from user-annotated acceleration data
    Bao, L
    Intille, SS
    [J]. PERVASIVE COMPUTING, PROCEEDINGS, 2004, 3001 : 1 - 17
  • [8] HuMAn: Complex Activity Recognition with Multi-Modal Multi-Positional Body Sensing
    Bharti, Pratool
    De, Debraj
    Chellappan, Sriram
    Das, Sajal K.
    [J]. IEEE TRANSACTIONS ON MOBILE COMPUTING, 2019, 18 (04) : 857 - 870
  • [9] Optimal classifier for imbalanced data using Matthews Correlation Coefficient metric
    Boughorbel, Sabri
    Jarray, Fethi
    El-Anbari, Mohammed
    [J]. PLOS ONE, 2017, 12 (06):
  • [10] GCHAR: An efficient Group-based Context-aware human activity recognition on smartphone
    Cao, Liang
    Wang, Yufeng
    Zhang, Bo
    Jin, Qun
    Vasilakos, Athanasios V.
    [J]. JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2018, 118 : 67 - 80