Enhanced multimodal emotion recognition in healthcare analytics: A deep learning based model-level fusion approach

被引:3
|
作者
Islam, Md. Milon [1 ]
Nooruddin, Sheikh [1 ]
Karray, Fakhri [1 ,2 ]
Muhammad, Ghulam [3 ]
机构
[1] Univ Waterloo, Ctr Pattern Anal & Machine Intelligence, Dept Elect & Comp Engn, Waterloo, ON N2L 3G1, Canada
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] King Saud Univ, Coll Comp & Informat Sci, Dept Comp Engn, Riyadh 11543, Saudi Arabia
基金
加拿大自然科学与工程研究理事会;
关键词
Multimodal emotion recognition; Depthwise separable convolutional neural; networks; Bi-directional long short-term memory; Soft attention; Healthcare analytics; CLASSIFICATION;
D O I
10.1016/j.bspc.2024.106241
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Deep learning techniques have drawn considerable interest in emotion recognition due to recent technological developments in healthcare analytics. Automatic patient emotion recognition can assist healthcare analytics by providing feedback to the stakeholders of competent healthcare about the conditions of the patients and their satisfaction levels. In this paper, we propose a novel model -level fusion technique based on deep learning for enhanced emotion recognition from multimodal signals to monitor patients in connected healthcare. The representative visual features from the video signals are extracted through the Depthwise Separable Convolution Neural Network, and the optimized temporal attributes are derived from the multiple physiological data utilizing Bi-directional Long Short -Term Memory. A soft attention method fused the high multimodal features obtained from the two data modalities to retrieve the most significant features by focusing on emotionally salient parts of the features. We exploited two face detection methods, Histogram of Oriented Gradients and Convolutional Neural Network -based face detector (ResNet-34), to observe the effects of facial features on emotion recognition. Lastly, extensive experimental evaluations have been conducted using the widely used Bio Vid Emo DB multimodal dataset to verify the performance of the proposed architecture. Experimental results show that the developed fusion architecture improved the accuracy of emotion recognition from multimodal signals and outperformed the performance of both state-of-the-art techniques and baseline methods.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Audio-Visual Fusion Network Based on Conformer for Multimodal Emotion Recognition
    Guo, Peini
    Chen, Zhengyan
    Li, Yidi
    Liu, Hong
    ARTIFICIAL INTELLIGENCE, CICAI 2022, PT II, 2022, 13605 : 315 - 326
  • [42] Multimodal emotion recognition from facial expression and speech based on feature fusion
    Tang, Guichen
    Xie, Yue
    Li, Ke
    Liang, Ruiyu
    Zhao, Li
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (11) : 16359 - 16373
  • [43] Multimodal emotion recognition from facial expression and speech based on feature fusion
    Guichen Tang
    Yue Xie
    Ke Li
    Ruiyu Liang
    Li Zhao
    Multimedia Tools and Applications, 2023, 82 : 16359 - 16373
  • [44] Predictive analytics of complex healthcare systems using deep learning based disease diagnosis model
    Saeed, Muhammad Kashif
    Al Mazroa, Alanoud
    Alghamdi, Bandar M.
    Alallah, Fouad Shoie
    Alshareef, Abdulrhman
    Mahmud, Ahmed
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [45] Dual-level Deep Evidential Fusion: Integrating multimodal information for enhanced reliable decision-making in deep learning
    Shao, Zhimin
    Dou, Weibei
    Pan, Yu
    INFORMATION FUSION, 2024, 103
  • [46] Improved Multimodal Emotion Recognition for Better Game-Based Learning
    Bahreini, Kiavash
    Nadolski, Rob
    Westera, Wim
    GAMES AND LEARNING ALLIANCE, GALA 2014, 2015, 9221 : 107 - 120
  • [47] An Adaptive Framework of Multimodal Emotion Recognition Based on Collaborative Discriminative Learning
    Wang, Yadi
    Guo, Xiaoding
    Zhang, Yibo
    Ren, Yiyuan
    Huang, Wendi
    Liu, Zunyan
    Feng, Yuming
    Dai, Xiangguang
    Zhang, Wei
    Che, Hangjun
    2023 15TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE, ICACI, 2023,
  • [48] Dynamic Muscle Fatigue State Recognition Based on Deep Learning Fusion Model
    Liu, Jingxuan
    Tao, Qing
    Wu, Bin
    IEEE ACCESS, 2023, 11 : 95079 - 95091
  • [49] Multimodal Approach of Speech Emotion Recognition Using Multi-Level Multi-Head Fusion Attention-Based Recurrent Neural Network
    Ngoc-Huynh Ho
    Yang, Hyung-Jeong
    Kim, Soo-Hyung
    Lee, Gueesang
    IEEE ACCESS, 2020, 8 : 61672 - 61686
  • [50] Group Gated Fusion on Attention-based Bidirectional Alignment for Multimodal Emotion Recognition
    Liu, Pengfei
    Li, Kun
    Meng, Helen
    INTERSPEECH 2020, 2020, : 379 - 383