Enhanced multimodal emotion recognition in healthcare analytics: A deep learning based model-level fusion approach

被引:3
|
作者
Islam, Md. Milon [1 ]
Nooruddin, Sheikh [1 ]
Karray, Fakhri [1 ,2 ]
Muhammad, Ghulam [3 ]
机构
[1] Univ Waterloo, Ctr Pattern Anal & Machine Intelligence, Dept Elect & Comp Engn, Waterloo, ON N2L 3G1, Canada
[2] Mohamed bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
[3] King Saud Univ, Coll Comp & Informat Sci, Dept Comp Engn, Riyadh 11543, Saudi Arabia
基金
加拿大自然科学与工程研究理事会;
关键词
Multimodal emotion recognition; Depthwise separable convolutional neural; networks; Bi-directional long short-term memory; Soft attention; Healthcare analytics; CLASSIFICATION;
D O I
10.1016/j.bspc.2024.106241
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Deep learning techniques have drawn considerable interest in emotion recognition due to recent technological developments in healthcare analytics. Automatic patient emotion recognition can assist healthcare analytics by providing feedback to the stakeholders of competent healthcare about the conditions of the patients and their satisfaction levels. In this paper, we propose a novel model -level fusion technique based on deep learning for enhanced emotion recognition from multimodal signals to monitor patients in connected healthcare. The representative visual features from the video signals are extracted through the Depthwise Separable Convolution Neural Network, and the optimized temporal attributes are derived from the multiple physiological data utilizing Bi-directional Long Short -Term Memory. A soft attention method fused the high multimodal features obtained from the two data modalities to retrieve the most significant features by focusing on emotionally salient parts of the features. We exploited two face detection methods, Histogram of Oriented Gradients and Convolutional Neural Network -based face detector (ResNet-34), to observe the effects of facial features on emotion recognition. Lastly, extensive experimental evaluations have been conducted using the widely used Bio Vid Emo DB multimodal dataset to verify the performance of the proposed architecture. Experimental results show that the developed fusion architecture improved the accuracy of emotion recognition from multimodal signals and outperformed the performance of both state-of-the-art techniques and baseline methods.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Multimodal Emotion Recognition with Deep Learning: Advancements, challenges, and future directions
    Geetha, A., V
    Mala, T.
    Priyanka, D.
    Uma, E.
    INFORMATION FUSION, 2024, 105
  • [22] YYDeepCarc: Deep Learning-Powered Carcinogenicity Prediction Using Model-Level Representation
    Li, Ting
    Tong, Weida
    Roberts, Ruth
    Liu, Zhichao
    Thakkar, Shraddha
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2021, 4
  • [23] Multimodal emotion recognition model via hybrid model with improved feature level fusion on facial and EEG feature set
    Singh P.
    Tripathi M.K.
    Patil M.B.
    Shivendra
    Neelakantappa M.
    Multimedia Tools and Applications, 2025, 84 (1) : 1 - 36
  • [24] Efficient Multimodal Biometric Recognition for Secure Authentication Based on Deep Learning Approach
    Rajasekar, Vani
    Saracevic, Muzafer
    Hassaballah, Mahmoud
    Karabasevic, Darjan
    Stanujkic, Dragisa
    Zajmovic, Mahir
    Tariq, Usman
    Jayapaul, Premalatha
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2023, 32 (03)
  • [25] MULTIMODAL EMOTION RECOGNITION WITH CAPSULE GRAPH CONVOLUTIONAL BASED REPRESENTATION FUSION
    Liu, Jiaxing
    Chen, Sen
    Wang, Longbiao
    Liu, Zhilei
    Fu, Yahui
    Guo, Lili
    Dang, Jianwu
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6339 - 6343
  • [26] Cross-Subject Multimodal Emotion Recognition Based on Hybrid Fusion
    Cimtay, Yucel
    Ekmekcioglu, Erhan
    Caglar-Ozhan, Seyma
    IEEE ACCESS, 2020, 8 : 168865 - 168878
  • [27] Feature-Enhanced Multimodal Interaction model for emotion recognition in conversation
    Fu, Yanping
    Yan, Xiaoyuan
    Chen, Wei
    Zhang, Jun
    KNOWLEDGE-BASED SYSTEMS, 2025, 309
  • [28] IoT-based approach to multimodal music emotion recognition
    Zhao, Hanbing
    Jin, Ling
    ALEXANDRIA ENGINEERING JOURNAL, 2025, 113 : 19 - 31
  • [29] Healthcare entity recognition based on deep learning
    He, Qinlu
    Gao, Pengze
    Zhang, Fan
    Bian, Genqing
    Li, Zhen
    Wang, Zan
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (11) : 32739 - 32763
  • [30] Deep Auto-Encoders With Sequential Learning for Multimodal Dimensional Emotion Recognition
    Nguyen, Dung
    Nguyen, Duc Thanh
    Zeng, Rui
    Nguyen, Thanh Thi
    Tran, Son N.
    Nguyen, Thin
    Sridharan, Sridha
    Fookes, Clinton
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1313 - 1324