Cluster-Level Contrastive Learning for Emotion Recognition in Conversations

被引:28
作者
Yang, Kailai [1 ,2 ]
Zhang, Tianlin [1 ,2 ]
Alhuzali, Hassan [3 ]
Ananiadou, Sophia [1 ,2 ]
机构
[1] Univ Manchester, NaCTeM, Manchester M13 9PL, England
[2] Univ Manchester, Dept Comp Sci, Manchester M13 9PL, England
[3] Umm Al Qura Univ, Coll Comp & Informat Syst, Mecca 24382, Saudi Arabia
基金
英国生物技术与生命科学研究理事会;
关键词
Emotion recognition; Prototypes; Linguistics; Task analysis; Semantics; Training; Adaptation models; Cluster-level contrastive learning; emotion recognition in conversations; pre-trained knowledge adapters; valence-arousal-dominance; DIALOGUE; FRAMEWORK;
D O I
10.1109/TAFFC.2023.3243463
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A key challenge for Emotion Recognition in Conversations (ERC) is to distinguish semantically similar emotions. Some works utilise Supervised Contrastive Learning (SCL) which uses categorical emotion labels as supervision signals and contrasts in high-dimensional semantic space. However, categorical labels fail to provide quantitative information between emotions. ERC is also not equally dependent on all embedded features in the semantic space, which makes the high-dimensional SCL inefficient. To address these issues, we propose a novel low-dimensional Supervised Cluster-level Contrastive Learning (SCCL) method, which first reduces the high-dimensional SCL space to a three-dimensional affect representation space Valence-Arousal-Dominance (VAD), then performs cluster-level contrastive learning to incorporate measurable emotion prototypes. To help modelling the dialogue and enriching the context, we leverage the pre-trained knowledge adapters to infuse linguistic and factual knowledge. Experiments show that our method achieves new state-of-the-art results with 69.81% on IEMOCAP, 65.7% on MELD, and 62.51% on DailyDialog datasets. The analysis also proves that the VAD space is not only suitable for ERC but also interpretable, with VAD prototypes enhancing its performance and stabilising the training of SCCL. In addition, the pre-trained knowledge adapters benefit the performance of the utterance encoder and SCCL. Our code is available at: https://github.com/SteveKGYang/SCCLI
引用
收藏
页码:3269 / 3280
页数:12
相关论文
共 50 条
  • [31] DialoguePCN: Perception and Cognition Network for Emotion Recognition in Conversations
    Wu, Xiaolong
    Feng, Chang
    Xu, Mingxing
    Zheng, Thomas Fang
    Hamdulla, Askar
    [J]. IEEE ACCESS, 2023, 11 : 141251 - 141260
  • [32] Multitask, Multilabel, and Multidomain Learning With Convolutional Networks for Emotion Recognition
    Pons, Gerard
    Masip, David
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (06) : 4764 - 4771
  • [33] Class Hierarchy Aware Contrastive Feature Learning for Multigranularity SAR Target Recognition
    Wen, Zaidao
    Wang, Zikai
    Zhang, Jianting
    Lv, Yafei
    Wu, Qian
    [J]. IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2024, 60 (06) : 7962 - 7977
  • [34] CCL: Continual Contrastive Learning for LiDAR Place Recognition
    Cui, Jiafeng
    Chen, Xieyuanli
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (08) : 4433 - 4440
  • [35] Hierarchical Group-Level Emotion Recognition
    Fujii, Katsuya
    Sugimura, Daisuke
    Hamamoto, Takayuki
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2021, 23 (23) : 3892 - 3906
  • [36] MM-DFN: MULTIMODAL DYNAMIC FUSION NETWORK FOR EMOTION RECOGNITION IN CONVERSATIONS
    Hu, Dou
    Hou, Xiaolong
    Wei, Lingwei
    Jiang, Lianxin
    Mo, Yang
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7037 - 7041
  • [37] Non-Contrastive Self-Supervised Learning for Utterance-Level Information Extraction From Speech
    Cho, Jaejin
    Villalba, Jesus
    Moro-Velazquez, Laureano
    Dehak, Najim
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1284 - 1295
  • [38] Frame-Level Teacher-Student Learning With Data Privacy for EEG Emotion Recognition
    Gu, Tianhao
    Wang, Zhe
    Xu, Xinlei
    Li, Dongdong
    Yang, Hai
    Du, Wenli
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 11021 - 11028
  • [39] Multi-Level Graph Knowledge Contrastive Learning
    Yang, Haoran
    Wang, Yuhao
    Zhao, Xiangyu
    Chen, Hongxu
    Yin, Hongzhi
    Li, Qing
    Xu, Guandong
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 8829 - 8841
  • [40] Generalized Contrastive Partial Label Learning for Cross-Subject EEG-Based Emotion Recognition
    Li, Wei
    Fan, Lingmin
    Shao, Shitong
    Song, Aiguo
    [J]. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 11