Tunisian Sign Language Recognition and Translation Using Deep Learning

被引:0
作者
El Askri, Marah [1 ]
Basly, Hend [1 ]
Bchir, Riadh [2 ]
Zayene, Mohamed Amine [1 ]
Sayadi, Fatma Ezzahra [1 ]
机构
[1] Univ Sousse, Natl Engn Sch Sousse, Lab Networked Objects Control & Commun Syst NOCCS, Eniso, Sousse, Tunisia
[2] Univ Monastir, Fac Sci Monastir, Monastir, Tunisia
来源
INTELLIGENT SYSTEMS AND PATTERN RECOGNITION, ISPR 2024, PT III | 2025年 / 2305卷
关键词
Deep Learning; LSTM; Sign Language Recognition;
D O I
10.1007/978-3-031-82156-1_3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human activity recognition, a dynamic field within computer vision and machine learning, aims to accurately identify and interpret human actions or activities using video data. This paper delves deeper into the domain by concentrating on Sign Language (SL) as a unique and intricate form of human activity. Sign Language (SL) is a vital communication tool, used by individuals who are deaf or hard of hearing, facilitating their interaction and engagement with others in a visually expressive manner. A lot of research and projects in SL are being done, and it has been observed that most of them are done in American Sign Language (ASL) or Indian Sign Language (ISL), but no such progress has been carried out for the other SLs. Therefore, this paper aims to present a work done in Arabic sign language (ArSL), especially Tunisian SL. The main idea is to recognize Tunisian SL and then translate it into spoken language using artificial intelligence (AI) methods. The purpose of this work is to ease the communication between individuals who are deaf and those who are not. The proposed method uses holistic landmarks with sequence key points for Tunisian sign language and trains it with a Long Short-Term Memory (LSTM) model. The results show that the proposed model can be used for real-time SL estimation, providing a better interpretation method for the deaf community. As a result, our model achieved a validation accuracy of 98% with a cross-validation technique.
引用
收藏
页码:26 / 39
页数:14
相关论文
共 16 条
[1]  
[Anonymous], About us
[2]  
[Anonymous], US
[3]  
ASLLVD, about us
[4]   Spatiotemporal Self-Attention Mechanism Driven by 3D Pose to Guide RGB Cues for Daily Living Human Activity Recognition [J].
Basly, Hend ;
Zayene, Mohamed Amine ;
Sayadi, Fatma Ezahra .
JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2023, 109 (01)
[5]   Deep Learning Shape Trajectories for Isolated Word Sign Language Recognition [J].
Fakhfakh, Sana ;
Ben Jemaa, Yousra .
INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2022, 19 (04) :660-666
[6]  
Grimes G. J., 1983, US Patent, Patent No. [4414537, 4 414 537]
[7]  
Hisham E., 2022, ESMAANI: A Static and Dynamic Arabic Sign Language Recognition System Based on Machine and Deep Learning Models, DOI 978-1-6654-8237-0/22/$31.00<(c)>2022
[8]  
i6.informatik.rwth-aachen, RWTH-BOSTON dataset
[9]  
Li DX, 2020, IEEE WINT CONF APPL, P1448, DOI [10.1109/WACV45572.2020.9093512, 10.1109/wacv45572.2020.9093512]
[10]   Isolated Sign Language Recognition with Multi-scale Features using LSTM [J].
Mercanoglu Sincan, Ozge ;
Tur, Anil Osman ;
Yalim Keles, Hacer .
2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,