Continuous Sign Language Recognition with Correlation Network

被引:43
|
作者
Hu, Lianyu [1 ]
Gao, Liqing [1 ]
Liu, Zekang [1 ]
Feng, Wei [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.00249
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human body trajectories are a salient cue to identify actions in the video. Such body trajectories are mainly conveyed by hands and face across consecutive frames in sign language. However, current methods in continuous sign language recognition (CSLR) usually process frames independently, thus failing to capture cross-frame trajectories to effectively identify a sign. To handle this limitation, we propose correlation network (CorrNet) to explicitly capture and leverage body trajectories across frames to identify signs. In specific, a correlation module is first proposed to dynamically compute correlation maps between the current frame and adjacent frames to identify trajectories of all spatial patches. An identification module is then presented to dynamically emphasize the body trajectories within these correlation maps. As a result, the generated features are able to gain an overview of local temporal movements to identify a sign. Thanks to its special attention on body trajectories, CorrNet achieves new state-of-the-art accuracy on four large-scale datasets, i.e., PHOENIX14, PHOENIX14-T, CSL-Daily, and CSL. A comprehensive comparison with previous spatial-temporal reasoning methods verifies the effectiveness of CorrNet. Visualizations demonstrate the effects of CorrNet on emphasizing human body trajectories across adjacent frames.
引用
收藏
页码:2529 / 2539
页数:11
相关论文
共 50 条
  • [41] Rethinking the temporal downsampling paradigm for continuous sign language recognition
    Liu, Caifeng
    Hu, Lianyu
    MULTIMEDIA SYSTEMS, 2025, 31 (02)
  • [42] Continuous Indian Sign Language Gesture Recognition and Sentence Formation
    Tripathi, Kumud
    Baranwal, Neha
    Nandi, G. C.
    ELEVENTH INTERNATIONAL CONFERENCE ON COMMUNICATION NETWORKS, ICCN 2015/INDIA ELEVENTH INTERNATIONAL CONFERENCE ON DATA MINING AND WAREHOUSING, ICDMW 2015/NDIA ELEVENTH INTERNATIONAL CONFERENCE ON IMAGE AND SIGNAL PROCESSING, ICISP 2015, 2015, 54 : 523 - 531
  • [43] Collaborative Multilingual Continuous Sign Language Recognition: A Unified Framework
    Hu, Hezhen
    Pu, Junfu
    Zhou, Wengang
    Li, Houqiang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7559 - 7570
  • [44] Continuous Sign Language Recognition Based on CM-Transformer
    Ye K.
    Zhang S.
    Guo Q.
    Li H.
    Cui X.
    Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications, 2022, 45 (05): : 49 - 53and78
  • [45] Scalable frame resolution for efficient continuous sign language recognition
    Hu, Lianyu
    Gao, Liqing
    Liu, Zekang
    Feng, Wei
    PATTERN RECOGNITION, 2024, 145
  • [46] Understanding vision-based continuous sign language recognition
    Neena Aloysius
    M. Geetha
    Multimedia Tools and Applications, 2020, 79 : 22177 - 22209
  • [47] Continuous Chinese Sign Language Recognition with CNN-LSTM
    Yang, Su
    Zhu, Qing
    NINTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2017), 2017, 10420
  • [48] DYNAMIC PSEUDO LABEL DECODING FOR CONTINUOUS SIGN LANGUAGE RECOGNITION
    Zhou, Hao
    Zhou, Wengang
    Li, Houqiang
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1282 - 1287
  • [49] An approach based on 1D fully convolutional network for continuous sign language recognition and labeling
    Fei Wang
    Chen Li
    Chuan-wen Liu
    Zhen Zeng
    Ke Xu
    Jin-xiu Wu
    Neural Computing and Applications, 2022, 34 : 17921 - 17935
  • [50] An approach based on 1D fully convolutional network for continuous sign language recognition and labeling
    Wang, Fei
    Li, Chen
    Liu, Chuan-wen
    Zeng, Zhen
    Xu, Ke
    Wu, Jin-xiu
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (20): : 17921 - 17935