A novel signal channel attention network for multi-modal emotion recognition

被引:1
|
作者
Du, Ziang [1 ]
Ye, Xia [1 ]
Zhao, Pujie [1 ]
机构
[1] Xian Res Inst High Tech, Xian, Shaanxi, Peoples R China
来源
关键词
hypercomplex neural networks; physiological signals; attention fusion module; multi-modal fusion; emotion recognition;
D O I
10.3389/fnbot.2024.1442080
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Physiological signal recognition is crucial in emotion recognition, and recent advancements in multi-modal fusion have enabled the integration of various physiological signals for improved recognition tasks. However, current models for emotion recognition with hyper complex multi-modal signals face limitations due to fusion methods and insufficient attention mechanisms, preventing further enhancement in classification performance. To address these challenges, we propose a new model framework named Signal Channel Attention Network (SCA-Net), which comprises three main components: an encoder, an attention fusion module, and a decoder. In the attention fusion module, we developed five types of attention mechanisms inspired by existing research and performed comparative experiments using the public dataset MAHNOB-HCI. All of these experiments demonstrate the effectiveness of the attention module we addressed for our baseline model in improving both accuracy and F1 score metrics. We also conducted ablation experiments within the most effective attention fusion module to verify the benefits of multi-modal fusion. Additionally, we adjusted the training process for different attention fusion modules by employing varying early stopping parameters to prevent model overfitting.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Cross-modal dynamic convolution for multi-modal emotion recognition
    Wen, Huanglu
    You, Shaodi
    Fu, Ying
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2021, 78
  • [32] Study of emotion recognition based on fusion multi-modal bio-signal with SAE and LSTM recurrent neural network
    Li Y.-J.
    Huang J.-J.
    Wang H.-Y.
    Zhong N.
    Tongxin Xuebao/Journal on Communications, 2017, 38 (12): : 109 - 120
  • [33] Multi-modal Emotion Recognition with Temporal-Band Attention Based on LSTM-RNN
    Liu, Jiamin
    Su, Yuanqi
    Liu, Yuehu
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT I, 2018, 10735 : 194 - 204
  • [34] M2FNet: Multi-modal Fusion Network for Emotion Recognition in Conversation
    Chudasama, Vishal
    Kar, Purbayan
    Gudmalwar, Ashish
    Shah, Nirmesh
    Wasnik, Pankaj
    Onoe, Naoyuki
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 4651 - 4660
  • [35] A Deep GRU-BiLSTM Network for Multi-modal Emotion Recognition from Text
    Yacoubi, Ibtissem
    Ferjaoui, Radhia
    Ben Khalifa, Anouar
    2024 IEEE 7TH INTERNATIONAL CONFERENCE ON ADVANCED TECHNOLOGIES, SIGNAL AND IMAGE PROCESSING, ATSIP 2024, 2024, : 138 - 143
  • [36] SMIN: Semi-Supervised Multi-Modal Interaction Network for Conversational Emotion Recognition
    Lian, Zheng
    Liu, Bin
    Tao, Jianhua
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2415 - 2429
  • [37] Multi-modal Conditional Attention Fusion for Dimensional Emotion Prediction
    Chen, Shizhe
    Jin, Qin
    MM'16: PROCEEDINGS OF THE 2016 ACM MULTIMEDIA CONFERENCE, 2016, : 571 - 575
  • [38] Emotion classification with multi-modal physiological signals using multi-attention-based neural network
    Zou, Chengsheng
    Deng, Zhen
    He, Bingwei
    Yan, Maosong
    Wu, Jie
    Zhu, Zhaoju
    COGNITIVE COMPUTATION AND SYSTEMS, 2024, 6 (1-3) : 1 - 11
  • [39] A Modality-Enhanced Multi-Channel Attention Network for Multi-Modal Dialogue Summarization
    Lu, Ming
    Liu, Yang
    Zhang, Xiaoming
    APPLIED SCIENCES-BASEL, 2024, 14 (20):
  • [40] Spread Spectrum and Conventional Modulation Signal Recognition Method Based on Generative Adversarial Network and Multi-modal Attention Mechanism
    Wang H.
    Zhang R.
    Huang Y.
    Dianzi Yu Xinxi Xuebao/Journal of Electronics and Information Technology, 2024, 46 (04): : 1212 - 1221