Ethiopian sign language recognition using deep convolutional neural network

被引:13
作者
Abeje, Bekalu Tadele [1 ]
Salau, Ayodeji Olalekan [2 ,3 ]
Mengistu, Abreham Debasu [4 ]
Tamiru, Nigus Kefyalew [5 ]
机构
[1] Haramaya Univ, Dept Informat Technol, Coll Comp & Informat, Dire Dawa, Ethiopia
[2] Afe Babalola Univ, Dept Elect Elect & Comp Engn, Ado Ekiti, Nigeria
[3] Saveetha Inst Med & Tech Sci, Saveetha Sch Engn, Chennai, Tamil Nadu, India
[4] Bahir Dar Univ, Inst Technol, Dept Comp Sci, Bahir Dar, Ethiopia
[5] Debre Markos Univ, Inst Technol, Sch Elect & Comp Engn, Debre Markos, Ethiopia
关键词
Ethiopian sign language; Deep convolutional neural network; Amharic alphabet;
D O I
10.1007/s11042-022-12768-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, several technologies have been utilized to bridge the communication gap between persons who have hearing or speaking impairments and those who don't. This paper presents the development of a novel sign language recognition system which translates Ethiopian sign language (ETHSL) to Amharic alphabets using computer vision technology and Deep Convolutional Neural Network (CNN). The system accepts sign language images as input and gives Amharic text as the desired output. The proposed system comprises of three main stages which are: preprocessing, feature extraction, and recognition. The methodology employed involves data acquisition, preprocessing the acquired data, background normalization, image resizing, region of interest (ROI) identification, noise removal, brightness adjustment, and feature extraction, while Deep Convolutional Neural Network (CNN) was used for end-to-end classification. The data used in this study was acquired from students with hearing impairments at the Debre Markos Teaching College with an iPhone 6s phone which has a resolution of 3024 x 4020. The images are in JPEG file format and were collected in a controlled environment. The proposed system was implemented using Kera's (Tensorflow2.3.0 as backend) in python and tested using the image dataset collected from Debre Markos Teaching College graduating students of 2012. The results show that the running time was minimized by adjusting the images to a suitable size and color. In addition, the results show an improved recognition accuracy compared to previous works. The proposed model achieves 98.5% training, 95.59% validation, and 98.3% testing accuracy of recognition.
引用
收藏
页码:29027 / 29043
页数:17
相关论文
共 16 条
[1]  
Admasu Y. F., 2010, Proceedings 10th International Conference on Intelligent Systems Design and Applications (ISDA 2010), P995, DOI 10.1109/ISDA.2010.5687057
[2]  
Bantupalli K, 2018, IEEE INT CONF BIG DA, P4896, DOI 10.1109/BigData.2018.8622141
[3]  
Gimibi, 2014, RECOGNITION ISOLATED, P174
[4]  
Patel, 2018, INT J INNOV RES COMP, P9101, DOI [10.15680/IJIRCCE.2018, DOI 10.15680/IJIRCCE.2018]
[5]  
Rao GA, 2018, 2018 CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION ENGINEERING SYSTEMS (SPACES), P194, DOI 10.1109/SPACES.2018.8316344
[6]  
Ravikiran J., 2010, Lecture Notes in Electrical Engineering, V52, P321, DOI [10.1007/978-90-481-3517-2-25, DOI 10.1007/978-90-481-3517-2-25]
[7]   American Sign Language Alphabet Recognition Using a Neuromorphic Sensor and an Artificial Neural Network [J].
Rivera-Acosta, Miguel ;
Ortega-Cisneros, Susana ;
Rivera, Jorge ;
Sandoval-Ibarra, Federico .
SENSORS, 2017, 17 (10)
[8]  
Salau Ayodeji Olalekan, 2019, 2019 International Conference on Signal Processing and Communication (ICSC), P158
[9]  
Singha J., 2013, International Journal of Computer Applications, V70, P17, DOI DOI 10.5120/12174-7306
[10]   Recognition of Amharic sign language with Amharic alphabet signs using ANN and SVM [J].
Tamiru, Nigus Kefyalew ;
Tekeba, Menore ;
Salau, Ayodeji Olalekan .
VISUAL COMPUTER, 2022, 38 (05) :1703-1718