DIDA: Dynamic Individual-to-integrateD Augmentation for Self-supervised Skeleton-Based Action Recognition

被引:0
|
作者
Hu, Haobo [1 ]
Li, Jianan [1 ]
Fan, Hongbin [2 ]
Zhao, Zhifu [2 ]
Zhou, Yangtao [1 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian, Peoples R China
[2] Xidian Univ, Sch Artificial Intelligence Engn, Xian, Shaanxi, Peoples R China
来源
PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VII | 2025年 / 15037卷
基金
中国国家自然科学基金;
关键词
Self-supervised skeleton-based action recognition; Contrastive learning; Dynamic augmentation; Intensity adjustment;
D O I
10.1007/978-981-97-8511-7_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised action recognition plays a crucial role by enabling machines to understand and interpret human actions without the need for numerous human-annotated labels. Contrastive learning, which compels the model to focus on discriminative features by constructing positive and negative sample pairs, is a highly effective method for achieving self-supervised action recognition. In contrastive learning, existing models focus on designing various augmentation methods and simply applying a fixed combination of these augmentations to generate the sample pairs. Nevertheless, there are two primary concerns associated with these methods: (1) The contentious strong augmentation could distort the structure of skeleton data and lead to semantic distortion. (2) Existing methods often apply augmentations uniformly, ignoring the unique characteristics of each augmentation technique. To address these problems, we propose the Dynamic Individual-to-integrateD Augmentation (DIDA) framework. This framework is designed with an innovative dual-phase structure. In the first phase, a close-loop feedback structure is applied to handle each augmentation separately and adjust their intensities dynamically based on immediate results. In the second phase, individual-to-integrated augmentation strategy with multi-level contrastive learning is designed to further enhance the feature representation ability of the model. Extensive experiments show that the proposed DIDA outperforms current state-of-the-art methods on the NTU60 and NTU120 datasets.
引用
收藏
页码:496 / 510
页数:15
相关论文
共 50 条
  • [1] Global and Local Contrastive Learning for Self-Supervised Skeleton-Based Action Recognition
    Hu, Jinhua
    Hou, Yonghong
    Guo, Zihui
    Gao, Jiajun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 10578 - 10589
  • [2] Modeling the Relative Visual Tempo for Self-supervised Skeleton-based Action Recognition
    Zhu, Yisheng
    Han, Hu
    Yu, Zhengtao
    Liu, Guangcan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 13867 - 13876
  • [3] A puzzle questions form training for self-supervised skeleton-based action recognition
    Moutik, Oumaima
    Sekkat, Hiba
    Tchakoucht, Taha Ait
    El Kari, Badr
    Alaoui, Ahmed El Hilali
    IMAGE AND VISION COMPUTING, 2024, 148
  • [4] Cross-stream contrastive learning for self-supervised skeleton-based action recognition
    Li, Ding
    Tang, Yongqiang
    Zhang, Zhizhong
    Zhang, Wensheng
    IMAGE AND VISION COMPUTING, 2023, 135
  • [5] Contrast-Reconstruction Representation Learning for Self-Supervised Skeleton-Based Action Recognition
    Wang, Peng
    Wen, Jun
    Si, Chenyang
    Qian, Yuntao
    Wang, Liang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6224 - 6238
  • [6] Skeleton-based Self-Supervised Feature Extraction for Improved Dynamic Hand Gesture Recognition
    Ikne, Omar
    Allaert, Benjamin
    Wannous, Hazem
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [7] Focalized contrastive view-invariant learning for self-supervised skeleton-based action recognition
    Men, Qianhui
    Ho, Edmond S. L.
    Shum, Hubert P. H.
    Leung, Howard
    NEUROCOMPUTING, 2023, 537 : 198 - 209
  • [8] DMMG: Dual Min-Max Games for Self-Supervised Skeleton-Based Action Recognition
    Guan, Shannan
    Yu, Xin
    Huang, Wei
    Fang, Gengfa
    Lu, Haiyan
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 395 - 407
  • [9] Multi-scale motion contrastive learning for self-supervised skeleton-based action recognition
    Wu, Yushan
    Xu, Zengmin
    Yuan, Mengwei
    Tang, Tianchi
    Meng, Ruxing
    Wang, Zhongyuan
    MULTIMEDIA SYSTEMS, 2024, 30 (05)
  • [10] Contrastive Mask Learning for Self-Supervised 3D Skeleton-Based Action Recognition
    Zhang, Haoyuan
    SENSORS, 2025, 25 (05)