Disentangled Variational Autoencoder for Emotion Recognition in Conversations

被引:5
|
作者
Yang, Kailai [1 ]
Zhang, Tianlin [1 ]
Ananiadou, Sophia [1 ]
机构
[1] Univ Manchester, Dept Comp Sci, NaCTeM, Manchester M13 9PL, England
基金
英国生物技术与生命科学研究理事会;
关键词
Task analysis; Emotion recognition; Hidden Markov models; Context modeling; Decoding; Oral communication; Gaussian distribution; Emotion recognition in conversations; variational autoencoder; valence-arousal-dominance; disentangled representations; DIALOGUE;
D O I
10.1109/TAFFC.2023.3280038
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In Emotion Recognition in Conversations (ERC), the emotions of target utterances are closely dependent on their context. Therefore, existing works train the model to generate the response of the target utterance, which aims to recognise emotions leveraging contextual information. However, adjacent response generation ignores long-range dependencies and provides limited affective information in many cases. In addition, most ERC models learn a unified distributed representation for each utterance, which lacks interpretability and robustness. To address these issues, we propose a VAD-disentangled Variational AutoEncoder (VAD-VAE), which first introduces a target utterance reconstruction task based on Variational Autoencoder, then disentangles three affect representations Valence-Arousal-Dominance (VAD) from the latent space. We also enhance the disentangled representations by introducing VAD supervision signals from a sentiment lexicon and minimising the mutual information between VAD distributions. Experiments show that VAD-VAE outperforms the state-of-the-art model on two datasets. Further analysis proves the effectiveness of each proposed module and the quality of disentangled VAD representations.
引用
收藏
页码:508 / 518
页数:11
相关论文
共 50 条
  • [1] Cluster-Level Contrastive Learning for Emotion Recognition in Conversations
    Yang, Kailai
    Zhang, Tianlin
    Alhuzali, Hassan
    Ananiadou, Sophia
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (04) : 3269 - 3280
  • [2] Disentangled Variational Autoencoder for Social Recommendation
    Zhang, Yongshuai
    Huang, Jiajin
    Yang, Jian
    NEURAL PROCESSING LETTERS, 2024, 56 (03)
  • [3] Variational Autoencoder based Latent Factor Decoding of Multichannel EEG for Emotion Recognition
    Li, Xiang
    Zhao, Zhigang
    Song, Dawei
    Zhang, Yazhou
    Niu, Chunyang
    Zhang, Junwei
    Huo, Jidong
    Li, Jing
    2019 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2019, : 684 - 687
  • [4] Facial landmark disentangled network with variational autoencoder
    Sen Liang
    Zhi-ze Zhou
    Yu-dong Guo
    Xuan Gao
    Ju-yong Zhang
    Hu-jun Bao
    Applied Mathematics-A Journal of Chinese Universities, 2022, 37 : 290 - 305
  • [5] Facial landmark disentangled network with variational autoencoder
    Liang, Sen
    Zhou, Zhi-ze
    Guo, Yu-dong
    Gao, Xuan
    Zhang, Ju-yong
    Bao, Hu-jun
    APPLIED MATHEMATICS-A JOURNAL OF CHINESE UNIVERSITIES SERIES B, 2022, 37 (02) : 290 - 305
  • [6] A Disentangled Recurrent Variational Autoencoder for Speech Enhancement
    Yan, Hegen
    Lu, Zhihua
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1697 - 1702
  • [7] DualVAE: Dual Disentangled Variational AutoEncoder for Recommendation
    Guo, Zhiqiang
    Li, Guohui
    Li, Jianjun
    Wang, Chaoyang
    Shi, Si
    PROCEEDINGS OF THE 2024 SIAM INTERNATIONAL CONFERENCE ON DATA MINING, SDM, 2024, : 571 - 579
  • [8] Autoencoder With Emotion Embedding for Speech Emotion Recognition
    Zhang, Chenghao
    Xue, Lei
    IEEE ACCESS, 2021, 9 : 51231 - 51241
  • [9] Reconstructing Neutral Face Expressions with Disentangled Variational Autoencoder
    Wiem, Grina
    Ali, Douik
    ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT II, 2024, 14496 : 83 - 94
  • [10] A Transformer-Based Model With Self-Distillation for Multimodal Emotion Recognition in Conversations
    Ma, Hui
    Wang, Jian
    Lin, Hongfei
    Zhang, Bo
    Zhang, Yijia
    Xu, Bo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 776 - 788