Efficient dense labelling of human activity sequences from wearables using fully convolutional networks

被引:76
作者
Yao, Rui [1 ,2 ]
Lin, Guosheng [3 ]
Shi, Qinfeng [2 ]
Ranasinghe, Damith C. [2 ]
机构
[1] China Univ Min & Technol, Sch Comp Sci & Technol, Xuzhou 221116, Peoples R China
[2] Univ Adelaide, Sch Comp Sci, Adelaide, SA 5005, Australia
[3] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
基金
中国博士后科学基金; 中国国家自然科学基金;
关键词
Human activity recognition; Time series sequence classification; Fully convolutional networks; RECOGNITION; SENSORS;
D O I
10.1016/j.patcog.2017.12.024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recognising human activities in sequential data from sensors is a challenging research area. A significant problem arises from the need to determine fixed sequence partitions (windows) to overcome the inability of a single sample to provide adequate information-about an activity; commonly overcome by using a fixed size sliding window over consecutive samples to extract information-either handcrafted or learned features-and predicting a single label for all the samples in the window. Two key issues arise from this approach: (i) the samples in one window may not always share the same label, a problem more significant for short duration activities such as gestures. We refer to this as the multi-class windows problem. (ii) the inferencing phase is constrained by the window size selected during training while the best window size is difficult to tune in practice. We propose an efficient method for predicting the label of each sample, which we call dense labelling, in a sequence of activity data of arbitrary length based on a fully convolutional network (FCN) design. In particular, our approach overcomes the problems posed by multi-class windows and fixed size sequence partitions imposed during training. Further, our network learns both features and the classifier automatically. We conduct extensive experiments and demonstrate that our proposed approach is able to outperform the state-of-the-arts in terms of sample-based classification and activity-based label misalignment measures on three challenging datasets: Opportunity, Hand Gesture, and our new dataset-an activity dataset we release based on a wearable sensor worn by hospitalised patients. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:252 / 266
页数:15
相关论文
共 38 条
[11]  
Guan Y., 2017, P ACM INTERACT MOB W, V1
[12]   Human action recognition using genetic algorithms and convolutional neural networks [J].
Ijjina, Earnest Paul ;
Chalavadi, Krishna Mohan .
PATTERN RECOGNITION, 2016, 59 :199-212
[13]   Gesture spotting with body-worn inertial sensors to detect user activities [J].
Junker, Holger ;
Amft, Oliver ;
Lukowicz, Paul ;
Troester, Gerhard .
PATTERN RECOGNITION, 2008, 41 (06) :2010-2024
[14]   ImageNet Classification with Deep Convolutional Neural Networks [J].
Krizhevsky, Alex ;
Sutskever, Ilya ;
Hinton, Geoffrey E. .
COMMUNICATIONS OF THE ACM, 2017, 60 (06) :84-90
[15]   DeepEar: Robust Smartphone Audio Sensing in Unconstrained Acoustic Environments using Deep Learning [J].
Lane, Nicholas D. ;
Georgiev, Petko ;
Qendro, Lorena .
PROCEEDINGS OF THE 2015 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING (UBICOMP 2015), 2015, :283-294
[16]   Gradient-based learning applied to document recognition [J].
Lecun, Y ;
Bottou, L ;
Bengio, Y ;
Haffner, P .
PROCEEDINGS OF THE IEEE, 1998, 86 (11) :2278-2324
[17]  
Long J, 2015, PROC CVPR IEEE, P3431, DOI 10.1109/CVPR.2015.7298965
[18]  
Nair V, 2010, ICML, DOI DOI 10.5555/3104322.3104425
[19]   Learning Human Identity From Motion Patterns [J].
Neverova, Natalia ;
Wolf, Christian ;
Lacey, Griffin ;
Fridman, Lex ;
Chandra, Deepak ;
Barbello, Brandon ;
Taylor, Graham .
IEEE ACCESS, 2016, 4 :1810-1820
[20]   Layered representations for human activity recognition [J].
Oliver, N ;
Horvitz, E ;
Garg, A .
FOURTH IEEE INTERNATIONAL CONFERENCE ON MULTIMODAL INTERFACES, PROCEEDINGS, 2002, :3-8