HMGAN: A Hierarchical Multi-Modal Generative Adversarial Network Model for Wearable Human Activity Recognition

被引:13
作者
Chen, Ling [1 ,2 ]
Hu, Rong [1 ,3 ]
Wu, Menghan [1 ,3 ]
Zhou, Xin [1 ,4 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci & Technol, Alibab Zhejiang Univ Joint Res Inst Frontier Tech, 38 Zheda Rd, Hangzhou 310027, Peoples R China
[3] Zhejiang Univ, Coll Comp Sci & Technol, 38 Zheda Rd, Hangzhou 310027, Peoples R China
[4] Zhejiang Univ, Sch Software Technol, 38 Zheda Rd, Hangzhou 310027, Peoples R China
来源
PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT | 2023年 / 7卷 / 03期
关键词
Wearable human activity recognition; multi-modal; generative adversarial network; VITAL SIGN; ACCELERATION;
D O I
10.1145/3610909
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Wearable Human Activity Recognition (WHAR) is an important research field of ubiquitous and mobile computing. Deep WHAR models suffer from the overfitting problem caused by the lack of a large amount and variety of labeled data, which is usually addressed by generating data to enlarge the training set, i.e., Data Augmentation (DA). Generative Adversarial Networks (GANs) have shown their excellent data generation ability, and the generalization ability of a classification model can be improved by GAN-based DA. However, existing GANs cannot make full use of the important modality information and fail to balance modality details and global consistency, which cannot meet the requirements of deep multi-modal WHAR. In this paper, a hierarchical multi-modal GAN model (HMGAN) is proposed for WHAR. HMGAN consists of multiple modal generators, one hierarchical discriminator, and one auxiliary classifier. Multiple modal generators can learn the complex multi-modal data distributions of sensor data. Hierarchical discriminator can provide discrimination outputs for both low-level modal discrimination losses and high-level overall discrimination loss to draw a balance between modality details and global consistency. Experiments on five public WHAR datasets demonstrate that HMGAN achieves the state-of-the-art performance for WHAR, outperforming the best baseline by an average of 3.4%, 3.8%, and 3.5% in accuracy, macro F1 score, and weighted F1 score, respectively.
引用
收藏
页数:27
相关论文
共 94 条
[71]   Hierarchical Complex Activity Representation and Recognition Using Topic Model and Classifier Level Fusion [J].
Peng, Liangying ;
Chen, Ling ;
Wu, Xiaojie ;
Guo, Haodong ;
Chen, Gencai .
IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2017, 64 (06) :1369-1379
[72]   Towards Multimodal Deep Learning for Activity Recognition on Mobile Devices [J].
Radu, Valentin ;
Lane, Nicholas D. ;
Bhattacharya, Sourav ;
Mascolo, Cecilia ;
Marina, Mahesh K. ;
Kawsar, Fahim .
UBICOMP'16 ADJUNCT: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING, 2016, :185-188
[73]   Uncovering Human Multimodal Activity Recognition with a Deep Learning Approach [J].
Ranieri, Caetano M. ;
Vargas, Patricia A. ;
Romero, Roseli A. F. .
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
[74]   Introducing a New Benchmarked Dataset for Activity Monitoring [J].
Reiss, Attila ;
Stricker, Didier .
2012 16TH INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS (ISWC), 2012, :108-109
[75]  
Roggen D., 2010, 2010 Seventh International Conference on Networked Sensing Systems (INSS 2010), P233, DOI 10.1109/INSS.2010.5573462
[76]   Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks [J].
Sandfort, Veit ;
Yan, Ke ;
Pickhardt, Perry J. ;
Summers, Ronald M. .
SCIENTIFIC REPORTS, 2019, 9 (1)
[77]   How Good Is My GAN? [J].
Shmelkov, Konstantin ;
Schmid, Cordelia ;
Alahari, Karteek .
COMPUTER VISION - ECCV 2018, PT II, 2018, 11206 :218-234
[78]   GAN-based Style Transformation to Improve Gesture-recognition Accuracy [J].
Suzuki, Noeru ;
Watanabe, Yuki ;
Nakazawa, Atsushi .
PROCEEDINGS OF THE ACM ON INTERACTIVE MOBILE WEARABLE AND UBIQUITOUS TECHNOLOGIES-IMWUT, 2020, 4 (04)
[79]  
Sztyler T, 2016, INT CONF PERVAS COMP
[80]   Profile to frontal face recognition in the wild using coupled conditional generative adversarial network [J].
Taherkhani, Fariborz ;
Talreja, Veeru ;
Dawson, Jeremy ;
Valenti, Matthew C. ;
Nasrabadi, Nasser M. .
IET BIOMETRICS, 2022, 11 (03) :260-276