Facial emotion recognition of deaf and hard-of-hearing students for engagement detection using deep learning

被引:21
作者
Lasri, Imane [1 ]
Riadsolh, Anouar [1 ]
Elbelkacemi, Mourad [1 ]
机构
[1] Mohammed V Univ Rabat, Fac Sci, Lab Concept & Syst Elect Signals & Informat, Rabat, Morocco
关键词
Facial emotion recognition; Deep convolutional neural networks; Transfer learning; Deafness; Student engagement; EXPRESSION RECOGNITION; FACE;
D O I
10.1007/s10639-022-11370-4
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Nowadays, facial expression recognition (FER) has drawn considerable attention from the research community in various application domains due to the recent advancement of deep learning. In the education field, facial expression recognition has the potential to evaluate students' engagement in a classroom environment, especially for deaf and hard-of-hearing students. Several works have been conducted on detecting students' engagement from facial expressions using traditional machine learning or convolutional neural network (CNN) with only a few layers. However, measuring deaf and hard-of-hearing students' engagement is yet an unexplored area for experimental research. Therefore, we propose in this study a novel approach for detecting the engagement level ('highly engaged', 'nominally engaged', and 'not engaged') from the facial emotions of deaf and hard-of-hearing students using a deep CNN (DCNN) model and transfer learning (TL) technique. A pre-trained VGG-16 model is employed and fine-tuned on the Japanese female facial expression (JAFFE) dataset and the Karolinska directed emotional faces (KDEF) dataset. Then, the performance of the proposed model is compared to seven different pre-trained DCNN models (VGG-19, Inception v3, DenseNet-121, DenseNet-169, MobileNet, ResNet-50, and Xception). On the 10-fold cross-validation case, the best-achieved test accuracies with VGG-16 are 98% and 99% on JAFFE and KDEF datasets, respectively. According to the obtained results, the proposed approach outperformed other state-of-the-art methods.
引用
收藏
页码:4069 / 4092
页数:24
相关论文
共 39 条
[1]  
Abadi Martin, 2016, Proceedings of OSDI '16: 12th USENIX Symposium on Operating Systems Design and Implementation. OSDI '16, P265
[2]  
Aifanti N., 2010, 11th International Workshop on Image Analysis for Multimedia Interactive Services WIAMIS 10, V10, P1
[3]  
[Anonymous], 2014, PROC CVPR IEEE, DOI DOI 10.1109/CVPR.2014.233
[4]  
[Anonymous], 2015, KERAS: The python deep learning library
[5]   Investigating the Impact of a Real-time, Multimodal Student Engagement Analytics Technology in Authentic Classrooms [J].
Aslan, Sinem ;
Alyuz, Nese ;
Tanriover, Cagri ;
Mete, Sinem E. ;
Okur, Eda ;
D'Mello, Sidney K. ;
Esme, Asli Arslan .
CHI 2019: PROCEEDINGS OF THE 2019 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2019,
[6]   A new ML-based approach to enhance student engagement in online environment [J].
Ayouni, Sarra ;
Hajjej, Fahima ;
Maddeh, Mohamed ;
Al-Otaibi, Shaha .
PLOS ONE, 2021, 16 (11)
[7]  
Bradski G, 2000, DR DOBBS J, V25, P120
[8]   Facial expressions of emotion (KDEF): Identification under different display-duration conditions [J].
Calvo, Manuel G. ;
Lundqvist, Daniel .
BEHAVIOR RESEARCH METHODS, 2008, 40 (01) :109-115
[9]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[10]  
Duchi J, 2011, J MACH LEARN RES, V12, P2121