Word-level Sign Language Recognition Using Linguistic Adaptation of 77 GHz FMCW Radar Data

被引:6
作者
Rahman, M. Mahbubur [1 ]
Mdrafi, Robiulhossain [2 ]
Gurbuz, Ali C. [2 ]
Malaia, Evie [3 ]
Crawford, Chris [4 ]
Griffin, Darrin [5 ]
Gurbuz, Sevgi Z. [1 ]
机构
[1] Univ Alabama, Dept Elect & Comp Engn, Tuscaloosa, AL 35487 USA
[2] Mississippi State Univ, Dept Elect & Comp Engn, Starkville, MS USA
[3] Univ Alabama, Dept Commun Disorders, Tuscaloosa, AL USA
[4] Univ Alabama, Dept Comp Sci, Tuscaloosa, AL 35487 USA
[5] Univ Alabama, Dept Commun Studies, Tuscaloosa, AL USA
来源
2021 IEEE RADAR CONFERENCE (RADARCONF21): RADAR ON THE MOVE | 2021年
基金
美国国家科学基金会;
关键词
ASL; sign language; gesture recognition; RF sensing; radar; micro-Doppler; deep learning;
D O I
10.1109/RadarConf2147009.2021.9455190
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Over the years, there has been much research in both wearable and video-based American Sign Language (ASL) recognition systems. However, the restrictive and invasive nature of these sensing modalities remains a significant disadvantage in the context of Deaf-centric smart environments or devices that are responsive to ASL. This paper investigates the efficacy of RF sensors for word-level ASL recognition in support of human-computer interfaces designed for deaf or hard-of-hearing individuals. A principal challenge is the training of deep neural networks given the difficulty in acquiring native ASL signing data. In this paper, adversarial domain adaptation is exploited to bridge the physical/kinematic differences between the copysigning of hearing individuals (repetition of sign motion after viewing a video), and native signing of Deaf individuals who are fluent in sign language. Domain adaptation results are compared with those attained by directly synthesizing ASL signs using generative adversarial networks (GANs). Kinematic improvements to the GAN architecture, such as the insertion of micro-Doppler signature envelopes in a secondary branch of the GAN, are utilized to boost performance. Word-level classification accuracy of 91.3% is achieved for 20 ASL words.
引用
收藏
页数:6
相关论文
共 38 条
[1]   Radar Signal Processing for Elderly Fall Detection The future for in-home monitoring [J].
Amin, Moeness G. ;
Zhang, Yimin D. ;
Ahmad, Fauzia ;
Ho, K. C. .
IEEE SIGNAL PROCESSING MAGAZINE, 2016, 33 (02) :71-80
[2]   TraVeLGAN: Image-to-image Translation by Transformation Vector Learning [J].
Amodio, Matthew ;
Krishnaswamy, Smita .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :8975-8984
[3]   Radar-Based Human Gait Recognition Using Dual-Channel Deep Convolutional Neural Network [J].
Bai, Xueru ;
Hui, Ye ;
Wang, Li ;
Zhou, Feng .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2019, 57 (12) :9767-9778
[4]   Hearing L2 Sign Language Learners: How Do They Perform on ASL Phonological Fluency? [J].
Beal, Jennifer S. ;
Faniel, Kia .
SIGN LANGUAGE STUDIES, 2018, 19 (02) :204-224
[5]  
Bhargava P, 2018, INTEL AI DEV PROGRAM
[6]   Motion characterization using optical flow and fractal complexity [J].
Borneman, Joshua D. ;
Malaia, Evie ;
Wilbur, Ronnie B. .
JOURNAL OF ELECTRONIC IMAGING, 2018, 27 (05)
[7]   Using multiple sensors for mobile sign language recognition [J].
Brashear, H ;
Starner, T ;
Lukowicz, P ;
Junker, H .
SEVENTH IEEE INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS, PROCEEDINGS, 2003, :45-52
[8]  
Chen V., 2019, The micro-Doppler effect in radar, V2nd
[9]  
Csoka F., 2017, IEEE SIGNAL PROCESSI
[10]  
Dong C, 2015, IEEE COMPUT SOC CONF