Deep learning-based classification with improved time resolution for physical activities of children

被引:11
|
作者
Jang, Yongwon [1 ,2 ]
Kim, Seunghwan [2 ]
Kim, Kiseong [1 ,3 ]
Lee, Doheon [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Bio & Brain Engn, Daejeon, South Korea
[2] Elect & Telecommun Res Inst, Biomed IT Res Dept, Daejeon, South Korea
[3] BioBrain Inc, Daejeon, South Korea
来源
PEERJ | 2018年 / 6卷
基金
新加坡国家研究基金会;
关键词
Physical activity; Children; Classification; Convolutional neural network; Time resolution; ACTIVITY RECOGNITION; OBESE ADULTS; ACCELEROMETER; MOTION; SENSORS;
D O I
10.7717/peerj.5764
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Background. The proportion of overweight and obese people has increased tremendously in a short period, culminating in a worldwide trend of obesity that is reaching epidemic proportions. Overweight and obesity are serious issues, especially with regard to children. This is because obese children have twice the risk of becoming obese as adults, as compared to non-obese children. Nowadays, many methods for maintaining a caloric balance exist; however, these methods are not applicable to children. In this study, a new approach for helping children monitor their activities using a convolutional neural network (CNN) is proposed, which is applicable for real-time scenarios requiring high accuracy. Methods. A total of 136 participants (86 boys and 50 girls), aged between 8.5 years and 12.5 years (mean 10.5, standard deviation 1.1), took part in this study. The participants performed various movement while wearing custom-made three-axis accelerometer modules around their waists. The data acquired by the accelerometer module was preprocessed by dividing them into small sets (128 sample points for 2.8 s). Approximately 183,600 data samples were used by the developed CNN for learning to classify ten physical activities : slow walking, fast walking, slow running, fast running, walking up the stairs, walking down the stairs, jumping rope, standing up, sitting down, and remaining still. Results. The developed CNN classified the ten activities with an overall accuracy of 81.2%. When similar activities were merged, leading to seven merged activities, the CNN classified activities with an overall accuracy of 91.1%. Activity merging also improved performance indicators, for the maximum case of 66.4% in recall, 48.5% in precision, and 57.4% in fl score . The developed CNN classifier was compared to conventional machine learning algorithms such as the support vector machine, decision tree, and k-nearest neighbor algorithms, and the proposed CNN classifier performed the best: CNN (81.2%) > SVM (64.8%) > DT (63.9%) > kNN (55.4%) (for ten activities); CNN (91.1%) > SVM (74.4%) > DT (73.2%) > kNN (65.3%) (for the merged seven activities). Discussion. The developed algorithm distinguished physical activities with improved time resolution using short-time acceleration signals from the physical activities performed by children. This study involved algorithm development, participant recruitment, IRB approval, custom-design of a data acquisition module, and data collection. The self-selected moving speeds for walking and running (slow and fast) and the structure of staircase degraded the performance of the algorithm. However, after similar activities were merged, the effects caused by the self-selection of speed were reduced. The experimental results show that the proposed algorithm performed better than conventional algorithms. Owing to its simplicity, the proposed algorithm could be applied to real-time applicaitons.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Detection and classification of white blood cells with an improved deep learning-based approach
    Akalin, Fatma
    Yumusak, Nejat
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2022, 30 (07) : 2725 - 2739
  • [2] Surface EMG signals and deep transfer learning-based physical action classification
    Demir, Fatih
    Bajaj, Varun
    Ince, Melih C.
    Taran, Sachin
    Sengur, Abdulkadir
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (12): : 8455 - 8462
  • [3] Surface EMG signals and deep transfer learning-based physical action classification
    Fatih Demir
    Varun Bajaj
    Melih C. Ince
    Sachin Taran
    Abdulkadir Şengür
    Neural Computing and Applications, 2019, 31 : 8455 - 8462
  • [4] Deep learning-based Cervical Cancer Classification
    Khoulqi, Ichrak
    Idrissi, Najlae
    2022 INTERNATIONAL CONFERENCE ON TECHNOLOGY INNOVATIONS FOR HEALTHCARE, ICTIH, 2022, : 30 - 33
  • [5] Deep Learning-Based Classification of Hyperspectral Data
    Chen, Yushi
    Lin, Zhouhan
    Zhao, Xing
    Wang, Gang
    Gu, Yanfeng
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2014, 7 (06) : 2094 - 2107
  • [6] Deep learning-based classification and segmentation for scalpels
    Su, Baiquan
    Zhang, Qingqian
    Gong, Yi
    Xiu, Wei
    Gao, Yang
    Xu, Lixin
    Li, Han
    Wang, Zehao
    Yu, Shi
    Hu, Yida David
    Yao, Wei
    Wang, Junchen
    Li, Changsheng
    Tang, Jie
    Gao, Li
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2023, 18 (05) : 855 - 864
  • [7] Deep learning-based software bug classification
    Meher, Jyoti Prakash
    Biswas, Sourav
    Mall, Rajib
    INFORMATION AND SOFTWARE TECHNOLOGY, 2024, 166
  • [8] Deep Learning-Based Classification of Diabetic Retinopathy
    Huang, Zhenjia
    PROCEEDINGS OF 2023 4TH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL INTELLIGENCE FOR MEDICINE SCIENCE, ISAIMS 2023, 2023, : 371 - 375
  • [9] Deep Learning-Based Water Crystal Classification
    Thi, Hien Doan
    Andres, Frederic
    Quoc, Long Tran
    Emoto, Hiro
    Hayashi, Michiko
    Katsumata, Ken
    Oshide, Takayuki
    APPLIED SCIENCES-BASEL, 2022, 12 (02):
  • [10] Deep learning-based classification and segmentation for scalpels
    Baiquan Su
    Qingqian Zhang
    Yi Gong
    Wei Xiu
    Yang Gao
    Lixin Xu
    Han Li
    Zehao Wang
    Shi Yu
    Yida David Hu
    Wei Yao
    Junchen Wang
    Changsheng Li
    Jie Tang
    Li Gao
    International Journal of Computer Assisted Radiology and Surgery, 2023, 18 : 855 - 864