Facial emotion recognition of deaf and hard-of-hearing students for engagement detection using deep learning

被引:18
作者
Lasri, Imane [1 ]
Riadsolh, Anouar [1 ]
Elbelkacemi, Mourad [1 ]
机构
[1] Mohammed V Univ Rabat, Fac Sci, Lab Concept & Syst Elect Signals & Informat, Rabat, Morocco
关键词
Facial emotion recognition; Deep convolutional neural networks; Transfer learning; Deafness; Student engagement; EXPRESSION RECOGNITION; FACE;
D O I
10.1007/s10639-022-11370-4
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Nowadays, facial expression recognition (FER) has drawn considerable attention from the research community in various application domains due to the recent advancement of deep learning. In the education field, facial expression recognition has the potential to evaluate students' engagement in a classroom environment, especially for deaf and hard-of-hearing students. Several works have been conducted on detecting students' engagement from facial expressions using traditional machine learning or convolutional neural network (CNN) with only a few layers. However, measuring deaf and hard-of-hearing students' engagement is yet an unexplored area for experimental research. Therefore, we propose in this study a novel approach for detecting the engagement level ('highly engaged', 'nominally engaged', and 'not engaged') from the facial emotions of deaf and hard-of-hearing students using a deep CNN (DCNN) model and transfer learning (TL) technique. A pre-trained VGG-16 model is employed and fine-tuned on the Japanese female facial expression (JAFFE) dataset and the Karolinska directed emotional faces (KDEF) dataset. Then, the performance of the proposed model is compared to seven different pre-trained DCNN models (VGG-19, Inception v3, DenseNet-121, DenseNet-169, MobileNet, ResNet-50, and Xception). On the 10-fold cross-validation case, the best-achieved test accuracies with VGG-16 are 98% and 99% on JAFFE and KDEF datasets, respectively. According to the obtained results, the proposed approach outperformed other state-of-the-art methods.
引用
收藏
页码:4069 / 4092
页数:24
相关论文
共 39 条
  • [1] Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
  • [2] Aifanti N., 2010, 11 INT WORKSH IM AN, P1
  • [3] [Anonymous], 2014, P IEEE C COMPUTER VI, DOI DOI 10.1109/CVPR.2014.233
  • [4] Investigating the Impact of a Real-time, Multimodal Student Engagement Analytics Technology in Authentic Classrooms
    Aslan, Sinem
    Alyuz, Nese
    Tanriover, Cagri
    Mete, Sinem E.
    Okur, Eda
    D'Mello, Sidney K.
    Esme, Asli Arslan
    [J]. CHI 2019: PROCEEDINGS OF THE 2019 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2019,
  • [5] A new ML-based approach to enhance student engagement in online environment
    Ayouni, Sarra
    Hajjej, Fahima
    Maddeh, Mohamed
    Al-Otaibi, Shaha
    [J]. PLOS ONE, 2021, 16 (11):
  • [6] Bradski G, 2000, DR DOBBS J, V25, P120
  • [7] Facial expressions of emotion (KDEF): Identification under different display-duration conditions
    Calvo, Manuel G.
    Lundqvist, Daniel
    [J]. BEHAVIOR RESEARCH METHODS, 2008, 40 (01) : 109 - 115
  • [8] Chollet F., 2015, Keras: The Python deep learning library
  • [9] Xception: Deep Learning with Depthwise Separable Convolutions
    Chollet, Francois
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1800 - 1807
  • [10] Duchi J, 2011, J MACH LEARN RES, V12, P2121