Recognition of Non-Manual Content in Continuous Japanese Sign Language

被引:14
作者
Brock, Heike [1 ]
Farag, Iva [2 ]
Nakadai, Kazuhiro [1 ]
机构
[1] Honda Res Inst Japan Co Ltd, Wako, Saitama 3510188, Japan
[2] Saarland Univ, Fac Sci & Engn, D-66123 Saarbrucken, Germany
关键词
sign language; learning systems; motion segmentation; signal processing; gesture information retrieval; neural networks;
D O I
10.3390/s20195621
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The quality of recognition systems for continuous utterances in signed languages could be largely advanced within the last years. However, research efforts often do not address specific linguistic features of signed languages, as e.g., non-manual expressions. In this work, we evaluate the potential of a single video camera-based recognition system with respect to the latter. For this, we introduce a two-stage pipeline based on two-dimensional body joint positions extracted from RGB camera data. The system first separates the data flow of a signed expression into meaningful word segments on the base of a frame-wise binary Random Forest. Next, every segment is transformed into image-like shape and classified with a Convolutional Neural Network. The proposed system is then evaluated on a data set of continuous sentence expressions in Japanese Sign Language with a variation of non-manual expressions. Exploring multiple variations of data representations and network parameters, we are able to distinguish word segments of specific non-manual intonations with 86% accuracy from the underlying body joint movement data. Full sentence predictions achieve a total Word Error Rate of 15.75%. This marks an improvement of 13.22% as compared to ground truth predictions obtained from labeling insensitive towards non-manual content. Consequently, our analysis constitutes an important contribution for a better understanding of mixed manual and non-manual content in signed communication.
引用
收藏
页码:1 / 21
页数:21
相关论文
共 41 条
[1]  
[Anonymous], 2009, Taiwan Sign Lang. Beyond
[2]   APOLOGIES - JAPANESE AND AMERICAN STYLES [J].
BARNLUND, DC ;
YOSHIOKA, M .
INTERNATIONAL JOURNAL OF INTERCULTURAL RELATIONS, 1990, 14 (02) :193-206
[3]  
Brock H, 2018, PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), P4247
[4]  
Brock H, 2018, IEEE ROMAN, P370, DOI 10.1109/ROMAN.2018.8525717
[5]   SubUNets: End-to-end Hand Shape and Continuous Sign Language Recognition [J].
Camgoz, Necati Cihan ;
Hadfield, Simon ;
Koller, Oscar ;
Bowden, Richard .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3075-3084
[6]   Neural Sign Language Translation [J].
Camgoz, Necati Cihan ;
Hadfield, Simon ;
Koller, Oscar ;
Ney, Hermann ;
Bowden, Richard .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7784-7793
[7]   OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields [J].
Cao, Zhe ;
Hidalgo, Gines ;
Simon, Tomas ;
Wei, Shih-En ;
Sheikh, Yaser .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (01) :172-186
[8]   Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields [J].
Cao, Zhe ;
Simon, Tomas ;
Wei, Shih-En ;
Sheikh, Yaser .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1302-1310
[9]   American Sign Language Recognition Using Leap Motion Sensor [J].
Chuan, Ching-Hua ;
Regina, Eric ;
Guardino, Caroline .
2014 13TH INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2014, :541-544
[10]   A Deep Neural Framework for Continuous Sign Language Recognition by Iterative Training [J].
Cui, Runpeng ;
Liu, Hu ;
Zhang, Changshui .
IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (07) :1880-1891