An Emotion Recognition Method Based on Eye Movement and Audiovisual Features in MOOC Learning Environment

被引:10
作者
Bao, Jindi [1 ]
Tao, Xiaomei [2 ,3 ]
Zhou, Yinghui [1 ]
机构
[1] Guilin Univ Technol, Sch Informat Sci & Engn, Guilin 541004, Peoples R China
[2] Guangxi Normal Univ, Sch Comp Sci & Engn, Guilin 541000, Peoples R China
[3] Guangxi Normal Univ, Sch Software, Guilin 541000, Peoples R China
基金
美国国家科学基金会;
关键词
Emotion recognition; feature extraction; massive online open course (MOOC); multimodal analysis;
D O I
10.1109/TCSS.2022.3221128
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, more and more people have begun to use massive online open course (MOOC) platforms for distance learning. However, due to the space-time isolation between teachers and students, the negative emotional state of students in MOOC learning cannot be identified timely. Therefore, students cannot receive immediate feedback about their emotional states. In order to identify and classify learners' emotions in video learning scenarios, we propose a multimodal emotion recognition method based on eye movement signals, audio signals, and video images. In this method, two novel features are proposed: feature of coordinate difference of eyemovement (FCDE) and pixel change rate sequence (PCRS). FCDE is extracted by combining eye movement coordinate trajectory and video optical flow trajectory, which can represent the learner's attention degree. PCRS is extracted from the video image, which can represent the speed of image switching. A feature extraction network based on convolutional neural network (CNN) (FE-CNN) is designed to extract the deep features of the three modals. The extracted deep features are inputted into the emotion classification CNN (EC-CNN) to classify the emotions, including interest, happiness, confusion, and boredom. In single modal identification, the recognition accuracies corresponding to the three modals are 64.32%, 74.67%, and 71.88%. The three modals are fused by feature-level fusion, decision-level fusion, and model-level fusion methods, and the evaluation experiment results show that the method of decision-level fusion achieved the highest score of 81.90% of emotion recognition. Finally, the effectiveness of FCDE, FE-CNN, and EC-CNN modules is verified by ablation experiments.
引用
收藏
页码:171 / 183
页数:13
相关论文
共 50 条
  • [21] Deep Learning Based Audio-Visual Emotion Recognition in a Smart Learning Environment
    Ivleva, Natalja
    Pentel, Avar
    Dunajeva, Olga
    Justsenko, Valeria
    TOWARDS A HYBRID, FLEXIBLE AND SOCIALLY ENGAGED HIGHER EDUCATION, VOL 1, ICL 2023, 2024, 899 : 420 - 431
  • [22] Emotion Recognition by Integrating Eye Movement Analysis and Facial Expression Model
    Thong Van Huynh
    Yang, Hyung-Jeong
    Lee, Guee-Sang
    Kim, Soo-Hyung
    Na, In-Seop
    PROCEEDINGS OF THE 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND SOFT COMPUTING (ICMLSC 2019), 2019, : 166 - 169
  • [23] Emotion recognition system for E-learning environment based on facial expressions
    Begum, Farzana
    Neelima, Arambam
    Valan, J. Arul
    SOFT COMPUTING, 2023, 27 (22) : 17257 - 17265
  • [24] Emotion recognition system for E-learning environment based on facial expressions
    Farzana Begum
    Arambam Neelima
    J. Arul Valan
    Soft Computing, 2023, 27 : 17257 - 17265
  • [25] Emotion recognition based on sparse representation of phase synchronization features
    Kong, Wanzeng
    Song, Xulin
    Sun, Junfeng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (14) : 21203 - 21217
  • [26] Music Emotion Recognition Based on Deep Learning: A Review
    Jiang, Xingguo
    Zhang, Yuchao
    Lin, Guojun
    Yu, Ling
    IEEE ACCESS, 2024, 12 : 157716 - 157745
  • [27] Research on athlete’s wrong movement prediction method based on multimodal eye movement recognition
    Wang L.
    International Journal of Reasoning-based Intelligent Systems, 2022, 14 (04) : 176 - 183
  • [28] The SCEEGNet: An Efficient Learning Method for Emotion Recognition Based on the Few Channels
    Zhang, Meng
    Ni, Shoudong
    Chen, Shanshan
    Wang, Xiaoan
    Luo, Junwen
    2023 7TH INTERNATIONAL CONFERENCE ON BIOMEDICAL ENGINEERING AND APPLICATIONS, ICBEA, 2023, : 23 - 28
  • [29] U-Shaped Distribution Guided Sign Language Emotion Recognition With Semantic and Movement Features
    Zhang, Jiangtao
    Wang, Qingshan
    Wang, Qi
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (04) : 2180 - 2191
  • [30] An Occam's Razor View on Learning Audiovisual Emotion Recognition with Small Training Sets
    Vielzeuf, Valentin
    Kervadec, Corentin
    Pateux, Stephane
    Lechervy, Alexis
    Jurie, Frederic
    ICMI'18: PROCEEDINGS OF THE 20TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2018, : 589 - 593