Fine-grained Detection of Academic Emotions with Spatial Temporal Graph Attention Networks using Facial Landmarks

被引:0
作者
Fwa, Hua Leong [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, Singapore, Singapore
来源
CSEDU: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED EDUCATION - VOL 2 | 2022年
关键词
Spatial; Temporal; Affective States; Facial Landmarks; Graph Attention Network; Gated Recurrent Unit; GAZE;
D O I
10.5220/0010921200003182
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
With the incidence of the Covid-19 pandemic, institutions have adopted online learning as the main lesson delivery channel. A common criticism of online learning is that sensing of learners' affective states such as engagement is lacking which degrades the quality of teaching. In this study, we propose automatic sensing of learners' affective states in an online setting with web cameras capturing their facial landmarks and head poses. We postulate that the sparsely connected facial landmarks can be modelled using a Graph Neural Network. Using the publicly available in the wild DAiSEE dataset, we modelled both the spatial and temporal dimensions of the facial videos with a deep learning architecture consisting of Graph Attention Networks and Gated Recurrent Units. The ablation study confirmed that the differencing of consecutive frames of facial landmarks and the addition of head poses enhance the detection performance. The results further demonstrated that the model performed well in comparison with other models and more importantly, is suited for implementation on mobile devices with its low computational requirements.
引用
收藏
页码:27 / 34
页数:8
相关论文
共 25 条
[1]   Improving state-of-the-art in Detecting Student Engagement with Resnet and TCN Hybrid Network [J].
Abedi, Ali ;
Khan, Shehroz S. .
2021 18TH CONFERENCE ON ROBOTS AND VISION (CRV 2021), 2021, :151-157
[2]   Optuna: A Next-generation Hyperparameter Optimization Framework [J].
Akiba, Takuya ;
Sano, Shotaro ;
Yanase, Toshihiko ;
Ohta, Takeru ;
Koyama, Masanori .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :2623-2631
[3]   Estimation of behavioral user state based on eye gaze and head pose-application in an e-learning environment [J].
Asteriadis, Stylianos ;
Tzouveli, Paraskevi ;
Karpouzis, Kostas ;
Kollias, Stefanos .
MULTIMEDIA TOOLS AND APPLICATIONS, 2009, 41 (03) :469-493
[4]   An Automated Optimal Engagement and Attention Detection System Using Electrocardiogram [J].
Belle, Ashwin ;
Hargraves, Rosalyn Hobson ;
Najarian, Kayvan .
COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE, 2012, 2012
[5]   EmotioNet: An accurate, real-time algorithm for the automatic annotation of a million facial expressions in the wild [J].
Benitez-Quiroz, C. Fabian ;
Srinivasan, Ramprakash ;
Martinez, Aleix M. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :5562-5570
[6]  
Cho K., 2014, CoRR, DOI DOI 10.3115/V1/W14-4012
[7]   Gaze tutor: A gaze-reactive intelligent tutoring system [J].
D'Mello, Sidney ;
Olney, Andrew ;
Williams, Claire ;
Hays, Patrick .
INTERNATIONAL JOURNAL OF HUMAN-COMPUTER STUDIES, 2012, 70 (05) :377-398
[8]  
Grafsgaard Joseph F., 2013, Artificial Intelligence in Education. Proceedings of 16th International Conference (AIED 2013): LNCS 7926, P1, DOI 10.1007/978-3-642-39112-5_1
[9]  
Guhan P., 2020, ABC NET SEMI SUPERVI
[10]  
Guo Xiaoxiao, 2019, P IEEE INT C COMP VI