Approaching the Real-World: Supporting Activity Recognition Training with Virtual IMU Data

被引:22
作者
Kwon, Hyeokhyen [1 ]
Wang, Bingyao [2 ]
Abowd, Gregory D. [3 ]
Ploetz, Thomas [1 ]
机构
[1] Georgia Inst Technol, Sch Interact Comp, Atlanta, GA 30332 USA
[2] Georgia Inst Technol, Coll Comp, Atlanta, GA 30332 USA
[3] Northeastern Univ, Dept Elect & Comp Engn, Boston, MA 02115 USA
来源
PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT | 2021年 / 5卷 / 03期
关键词
Activity Recognition; Data Collection; Machine Learning;
D O I
10.1145/3478096
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, IMUTube introduced a paradigm change for bootstrapping human activity recognition (HAR) systems for wearables. The key idea is to utilize videos of activities to support training activity recognizers based on inertial measurement units (IMUs). This system retrieves video from public repositories and subsequently generates virtual IMU data from this. The ultimate vision for such a system is to make large amounts of weakly labeled videos accessible for model training in HAR and, as such, to overcome one of the most pressing issues in the field: the lack of significant amounts of labeled sample data. In this paper we present the first in-detail exploration of IMUTube in a realistic assessment scenario: the analysis of free-weight gym exercises. We make significant progress towards a flexible, fully-functional IMUTube system by extending it such that it can handle a range of artifacts that are common in unrestricted online videos, including various forms of video noise, non-human poses, body part occlusions, and extreme camera and human motion. By overcoming these real-world challenges, we are able to generate high-quality virtual IMU data, which allows us to employ IMUTube for practical analysis tasks. We show that HAR systems trained by incorporating virtual sensor data generated by IMUTube significantly outperform baseline models trained only with real IMU data. In doing so we demonstrate the practical utility of IMUTube and the progress made towards the final vision of the new bootstrapping paradigm.
引用
收藏
页数:32
相关论文
共 119 条
  • [61] A simple yet effective baseline for 3d human pose estimation
    Martinez, Julieta
    Hossain, Rayat
    Romero, Javier
    Little, James J.
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2659 - 2668
  • [62] Using Deep Data Augmentation Training to Address Software and Hardware Heterogeneities in Wearable and Smartphone Sensing Devices
    Mathur, Akhil
    Zhang, Tianlin
    Bhattacharya, Sourav
    Velickovic, Petar
    Joffe, Leonid
    Lane, Nicholas D.
    Kawsar, Fahim
    Lio, Pietro
    [J]. 2018 17TH ACM/IEEE INTERNATIONAL CONFERENCE ON INFORMATION PROCESSING IN SENSOR NETWORKS (IPSN), 2018, : 200 - 211
  • [63] A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation
    Mayer, Nikolaus
    Ilg, Eddy
    Hausser, Philip
    Fischer, Philipp
    Cremers, Daniel
    Dosovitskiy, Alexey
    Brox, Thomas
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4040 - 4048
  • [64] XNect: Real-time Multi-Person 3D Motion Capture with a Single RGB Camera
    Mehta, Dushyant
    Sotnychenko, Oleksandr
    Mueller, Franziska
    Xu, Weipeng
    Elgharib, Mohamed
    Fua, Pascal
    Seidel, Hans-Peter
    Rhodin, Helge
    Pons-Moll, Gerard
    Theobalt, Christian
    [J]. ACM TRANSACTIONS ON GRAPHICS, 2020, 39 (04):
  • [65] Meister S, 2018, AAAI CONF ARTIF INTE, P7251
  • [66] RecoFit: Using a Wearable Sensor to Find, Recognize, and Count Repetitive Exercises
    Morris, Dan
    Saponas, T. Scott
    Guillory, Andrew
    Kelner, Ilya
    [J]. 32ND ANNUAL ACM CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI 2014), 2014, : 3225 - 3234
  • [67] Mounsaveng S., 2019, arXiv
  • [68] Stacked Hourglass Networks for Human Pose Estimation
    Newell, Alejandro
    Yang, Kaiyu
    Deng, Jia
    [J]. COMPUTER VISION - ECCV 2016, PT VIII, 2016, 9912 : 483 - 499
  • [69] Nie Xuecheng, 2018, EUROPEAN C COMPUTER
  • [70] Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition
    Ordonez, Francisco Javier
    Roggen, Daniel
    [J]. SENSORS, 2016, 16 (01)