DMMG: Dual Min-Max Games for Self-Supervised Skeleton-Based Action Recognition

被引:0
|
作者
Guan, Shannan [1 ]
Yu, Xin [2 ]
Huang, Wei [3 ]
Fang, Gengfa [4 ]
Lu, Haiyan [1 ]
机构
[1] Univ Technol Sydney, Australian Artificial Intelligence Inst, Ultimo, NSW 2007, Australia
[2] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld 4072, Australia
[3] RIKEN Ctr Adv Intelligence Project, Tokyo 1030027, Japan
[4] Univ Technol Sydney, Sch Elect & Data Engn, Ultimo, NSW 2007, Australia
关键词
Self-supervised learning; adversarial learning; contrastive learning; skeleton action recognition; min-max game;
D O I
10.1109/TIP.2023.3338410
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we propose a new Dual Min-Max Games (DMMG) based self-supervised skeleton action recognition method by augmenting unlabeled data in a contrastive learning framework. Our DMMG consists of a viewpoint variation min-max game and an edge perturbation min-max game. These two min-max games adopt an adversarial paradigm to perform data augmentation on the skeleton sequences and graph-structured body joints, respectively. Our viewpoint variation min-max game focuses on constructing various hard contrastive pairs by generating skeleton sequences from various viewpoints. These hard contrastive pairs help our model learn representative action features, thus facilitating model transfer to downstream tasks. Moreover, our edge perturbation min-max game specializes in building diverse hard contrastive samples through perturbing connectivity strength among graph-based body joints. The connectivity-strength varying contrastive pairs enable the model to capture minimal sufficient information of different actions, such as representative gestures for an action while preventing the model from overfitting. By fully exploiting the proposed DMMG, we can generate sufficient challenging contrastive pairs and thus achieve discriminative action feature representations from unlabeled skeleton data in a self-supervised manner. Extensive experiments demonstrate that our method achieves superior results under various evaluation protocols on widely-used NTU-RGB+D, NTU120-RGB+D and PKU-MMD datasets.
引用
收藏
页码:395 / 407
页数:13
相关论文
共 50 条
  • [1] Global and Local Contrastive Learning for Self-Supervised Skeleton-Based Action Recognition
    Hu, Jinhua
    Hou, Yonghong
    Guo, Zihui
    Gao, Jiajun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 10578 - 10589
  • [2] Modeling the Relative Visual Tempo for Self-supervised Skeleton-based Action Recognition
    Zhu, Yisheng
    Han, Hu
    Yu, Zhengtao
    Liu, Guangcan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 13867 - 13876
  • [3] A puzzle questions form training for self-supervised skeleton-based action recognition
    Moutik, Oumaima
    Sekkat, Hiba
    Tchakoucht, Taha Ait
    El Kari, Badr
    Alaoui, Ahmed El Hilali
    IMAGE AND VISION COMPUTING, 2024, 148
  • [4] Cross-stream contrastive learning for self-supervised skeleton-based action recognition
    Li, Ding
    Tang, Yongqiang
    Zhang, Zhizhong
    Zhang, Wensheng
    IMAGE AND VISION COMPUTING, 2023, 135
  • [5] Contrast-Reconstruction Representation Learning for Self-Supervised Skeleton-Based Action Recognition
    Wang, Peng
    Wen, Jun
    Si, Chenyang
    Qian, Yuntao
    Wang, Liang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6224 - 6238
  • [6] DIDA: Dynamic Individual-to-integrateD Augmentation for Self-supervised Skeleton-Based Action Recognition
    Hu, Haobo
    Li, Jianan
    Fan, Hongbin
    Zhao, Zhifu
    Zhou, Yangtao
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT VII, 2025, 15037 : 496 - 510
  • [7] Focalized contrastive view-invariant learning for self-supervised skeleton-based action recognition
    Men, Qianhui
    Ho, Edmond S. L.
    Shum, Hubert P. H.
    Leung, Howard
    NEUROCOMPUTING, 2023, 537 : 198 - 209
  • [8] Multi-scale motion contrastive learning for self-supervised skeleton-based action recognition
    Wu, Yushan
    Xu, Zengmin
    Yuan, Mengwei
    Tang, Tianchi
    Meng, Ruxing
    Wang, Zhongyuan
    MULTIMEDIA SYSTEMS, 2024, 30 (05)
  • [9] Contrastive Mask Learning for Self-Supervised 3D Skeleton-Based Action Recognition
    Zhang, Haoyuan
    SENSORS, 2025, 25 (05)
  • [10] Frequency Decoupled Masked Auto-Encoder for Self-Supervised Skeleton-Based Action Recognition
    Liu, Ye
    Shi, Tianhao
    Zhai, Mingliang
    Liu, Jun
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 546 - 550