Self-Supervised Learning for Complex Activity Recognition Through Motif Identification Learning

被引:0
作者
Xia, Qingxin [1 ,2 ]
Morales, Jaime [2 ]
Huang, Yongzhi [1 ]
Hara, Takahiro [2 ]
Wu, Kaishun [1 ]
Oshima, Hirotomo [3 ]
Fukuda, Masamitsu [3 ]
Namioka, Yasuo [3 ]
Maekawa, Takuya [2 ]
机构
[1] Hong Kong Univ Sci & Technol Guangzhou, Informat Hub, Guangzhou 511458, Peoples R China
[2] Osaka Univ, Informat Sci & Technol, Suita, Osaka 5650871, Japan
[3] Toshiba Co Ltd, Corp Mfg Engn Ctr, Kawasaki, Kanagawa 2350017, Japan
关键词
Activity recognition; industrial domain; self-supervised learning; wearable sensor;
D O I
10.1109/TMC.2024.3514736
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Owing to the cost of collecting labeled sensor data, self-supervised learning (SSL) methods for human activity recognition (HAR) that effectively use unlabeled data for pretraining have attracted attention. However, applying prior SSL to COMPLEX activities in real industrial settings poses challenges. Despite the consistency of work procedures, varying circumstances, such as different sizes of packages and contents in a packing process, introduce significant variability within the same activity class. In this study, we focus on sensor data corresponding to characteristic and necessary actions (sensor data motifs) in a specific activity such as a stretching packing tape action in an assembling a box activity, and propose to train a neural network in self-supervised learning so that it identifies occurrences of the characteristic actions, i.e., Motif Identification Learning (MoIL). The feature extractor in the network is subsequently employed in the downstream activity recognition task, enabling accurate recognition of activities containing these characteristic actions, even with limited labeled training data. The MoIL approach was evaluated on real-world industrial activity data, encompassing the state-of-the-art SSL tasks with an improvement of up to 23.85% under limited training labels.
引用
收藏
页码:3779 / 3793
页数:15
相关论文
共 55 条
  • [11] Feldhorst Sascha, 2016, Proceedings of the 5th International Conference on Pattern Recognition Applications and Methods (ICPRAM 2016), P706
  • [12] Multi-view stacking for activity recognition with sound and accelerometer data
    Garcia-Ceja, Enrique
    Galvan-Tejada, Carlos E.
    Brena, Ramon
    [J]. INFORMATION FUSION, 2018, 40 : 45 - 56
  • [13] Gidaris S., 2018, P INT C LEARN REPR A
  • [14] Knowledge Distillation: A Survey
    Gou, Jianping
    Yu, Baosheng
    Maybank, Stephen J.
    Tao, Dacheng
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2021, 129 (06) : 1789 - 1819
  • [15] Grill J-B., 2020, ADV NEURAL INFORM PR, V33, P21271
  • [16] Haresamudram H, 2020, IEEE INT SYM WRBL CO, P45, DOI 10.1145/3410531.3414306
  • [17] Assessing the State of Self-Supervised Human Activity Recognition Using Wearables
    Haresamudram, Harish
    Essa, Irfan
    Plotz, Thomas
    [J]. PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2022, 6 (03):
  • [18] On the Role of Features in Human Activity Recognition
    Haresamudram, Harish
    Anderson, David, V
    Plotz, Thomas
    [J]. ISWC'19: PROCEEDINGS OF THE 2019 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS, 2019, : 78 - 88
  • [19] Contrastive Predictive Coding for Human Activity Recognition
    Haresamudram, Harish
    Essa, Irfan
    Plotz, Thomas
    [J]. PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2021, 5 (02):
  • [20] Hernandex-Cruz N., 2020, SN Comput. Sci., V1, P66, DOI [10.1007/s42979-020-0070-4, DOI 10.1007/S42979-020-0070-4]