A Real-time skeleton-based fall detection algorithm based on temporal convolutional networks and transformer encoder

被引:0
|
作者
Yu, Xiaoqun [1 ]
Wang, Chenfeng [1 ]
Wu, Wenyu [1 ]
Xiong, Shuping [2 ]
机构
[1] Southeast Univ, Sch Mech Engn, Dept Mech & Ind Design, Nanjing 211189, Peoples R China
[2] Korea Adv Inst Sci & Technol KAIST, Dept Ind & Syst Engn, Daejeon 34141, South Korea
基金
新加坡国家研究基金会;
关键词
Aging; Fall detection; Pose estimation; Temporal convolutional network; Transformer; Edge computing; RECOGNITION;
D O I
10.1016/j.pmcj.2025.102016
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As the population of older individuals living independently rises, coupled with the heightened risk of falls among this demographic, the need for automatic fall detection systems becomes increasingly urgent to ensure timely medical intervention. Computer vision (CV)-based methodologies have emerged as a preferred approach among researchers due to their contactless and pervasive nature. However, existing CV-based solutions often suffer from either poor robustness or prohibitively high computational requirements, impeding their practical implementation in elderly living environments. To address these challenges, we introduce TCNTE, a real-time skeleton-based fall detection algorithm that combines Temporal Convolutional Network (TCN) with Transformer Encoder (TE). We also successfully mitigate the severe class imbalance issue by implementing weighted focal loss. Cross-validation on multiple publicly available vision-based fall datasets demonstrates TCNTE's superiority over individual models (TCN and TE) and existing state-of-the-art fall detection algorithms, achieving remarkable accuracies (front view of UPFall: 99.58 %; side view of UP-Fall: 98.75 %; Le2i: 97.01 %; GMDCSA-24: 92.99 %) alongside practical viability. Visualizations using t-distributed stochastic neighbor embedding (t-SNE) reveal TCNTE's superior separation margin and cohesive clustering between fall and non-fall classes compared to TCN and TE. Crucially, TCNTE is designed for pervasive deployment in mobile and resource-constrained environments. Integrated with YOLOv8 pose estimation and BoT-SORT human tracking, the algorithm operates on NVIDIA Jetson Orin NX edge device, achieving an average frame rate of 19 fps for single-person and 17 fps for two-person scenarios. With its validated accuracy and impressive real-time performance, TCNTE holds significant promise for practical fall detection applications in older adult care settings.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] STST: Spatial-Temporal Specialized Transformer for Skeleton-based Action Recognition
    Zhang, Yuhan
    Wu, Bo
    Li, Wen
    Duan, Lixin
    Gan, Chuang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3229 - 3237
  • [22] Local and Global Spatial-Temporal Transformer for skeleton-based action recognition
    Liu, Ruyi
    Chen, Yu
    Gai, Feiyu
    Liu, Yi
    Miao, Qiguang
    Wu, Shuai
    NEUROCOMPUTING, 2025, 634
  • [23] Real-time prediction of horizontal drilling pressure based on convolutional Transformer
    Yan, Baoyong
    Tian, Jialin
    Wan, Jun
    Qiu, Yu
    Chen, Weiming
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (10)
  • [24] Decoupled spatio-temporal grouping transformer for skeleton-based action recognition
    Sun, Shengkun
    Jia, Zihao
    Zhu, Yisheng
    Liu, Guangcan
    Yu, Zhengtao
    VISUAL COMPUTER, 2024, 40 (08) : 5733 - 5745
  • [25] A novel real-time fall detection method based on head segmentation and convolutional neural network
    Yao, Chenguang
    Hu, Jun
    Min, Weidong
    Deng, Zhifeng
    Zou, Song
    Min, Weiqiong
    JOURNAL OF REAL-TIME IMAGE PROCESSING, 2020, 17 (06) : 1939 - 1949
  • [26] Real-time fall detection algorithm based on FFD-AlphaPose and CTR–GCN
    Xuecun Yang
    Yixiang Wang
    Zhonghua Dong
    Jiayu Li
    Qingyun Zhang
    Shushan Qiang
    Journal of Real-Time Image Processing, 2025, 22 (3)
  • [27] Two-stream spatio-temporal GCN-transformer networks for skeleton-based action recognition
    Chen, Dong
    Chen, Mingdong
    Wu, Peisong
    Wu, Mengtao
    Zhang, Tao
    Li, Chuanqi
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [28] Multitarget Real-Time Tracking Algorithm Based on Transformer and BYTE
    Pan Hao
    Liu Xiang
    Zhao Jingwen
    Zhang Xing
    LASER & OPTOELECTRONICS PROGRESS, 2023, 60 (06)
  • [29] Variation-aware directed graph convolutional networks for skeleton-based action
    Li, Tianchen
    Geng, Pei
    Cai, Guohui
    Hou, Xinran
    Lu, Xuequan
    Lyu, Lei
    KNOWLEDGE-BASED SYSTEMS, 2024, 302
  • [30] Pose-Guided Graph Convolutional Networks for Skeleton-Based Action Recognition
    Chen, Han
    Jiang, Yifan
    Ko, Hanseok
    IEEE ACCESS, 2022, 10 : 111725 - 111731