Multi-label emotion classification in texts using transfer learning

被引:47
作者
Ameer, Iqra [1 ]
Bolucu, Necva [2 ]
Siddiqui, Muhammad Hammad Fahim [1 ]
Can, Burcu [3 ]
Sidorov, Grigori [1 ]
Gelbukh, Alexander [1 ]
机构
[1] Inst Politecn Nacl IPN, Ctr Invest Computat CIC, Mexico City, Cdmx, Mexico
[2] Hacettepe Univ, Dept Comp Engn, Ankara, Turkiye
[3] Univ Wolverhampton, Res Inst Informat & Language Proc, Wolverhampton, England
关键词
Multi-label emotion classification; Bi-LSTM; Transformer Networks; Attention mechanism; Social media; MODEL;
D O I
10.1016/j.eswa.2022.118534
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Social media is a widespread platform that provides a massive amount of user-generated content that can be mined to reveal the emotions of social media users. This has many potential benefits, such as getting a sense of people's pulse on various events or news. Emotion classification from social media posts is challenging, especially when it comes to detecting multiple emotions from a short piece of text, as in multi-label classification problem. Most of the previous work on emotion detection has focused on deep neural networks such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) such as Long Short -Term Memory (LSTM) Networks. However, none of them has utilized multiple attention mechanisms and Recurrent Neural Networks (i.e., specialized attention networks for each emotion) nor utilized the recently introduced Transformer Networks such as XLNet, DistilBERT, and RoBERTa for the task of classifying emotions with multiple labels. The proposed multiple attention mechanism reveals the contribution of each word on each emotion, which has not been investigated before. In this study, we investigate both the use of LSTMs and the fine-tuning of Transformer Networks through Transfer Learning along with a single-attention network and a multiple-attention network for multi-label emotion classification. The experimental results show that our novel transfer learning models using pre-trained transformers with and without multiple attention mechanisms were able to outperform the current state-of-the-art accuracy (58.8% -Baziotis et al., 2018) in the SemEval-2018 Task-1C dataset. Our best-performing RoBERTa-MA (RoBERTa-Multi-attention) model outperformed the state-of-the-art and achieved 62.4% accuracy (3.6% gain over the state-of-the-art) on the challenging SemEval-2018 E-c: Detecting Emotions (multi-label classification) dataset for English. Moreover, the XLNet-MA (XLNet-Multi-attention) model outperformed the other proposed models by achieving 45.6% accuracy on the Ren-CECps dataset for Chinese.
引用
收藏
页数:13
相关论文
共 67 条
  • [1] Abdou M, 2018, P 12 INT WORKSH SEM, P210
  • [2] A survey of state-of-the-art approaches for emotion recognition in text
    Alswaidan, Nourah
    Menai, Mohamed El Bachir
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2020, 62 (08) : 2937 - 2987
  • [3] Ameer I, 2022, Arxiv, DOI arXiv:2207.01012
  • [4] Multi-Label Emotion Classification on Code-Mixed Text: Data and Methods
    Ameer, Iqra
    Sidorov, Grigori
    Gomez-Adorno, Helena
    Nawab, Rao Muhammad Adeel
    [J]. IEEE ACCESS, 2022, 10 : 8779 - 8789
  • [5] Ameer I, 2020, COMPUT SIST, V24, P1159, DOI [10.13053/CyS-24-3-3476, 10.13053/cys-24-3-3476]
  • [6] Apidianaki M., 2018, P 12 INT WORKSHOP SE, P172, DOI [10.18653/ v1/s18-1025, DOI 10.18653/V1/S18-1025]
  • [7] Asghar MZ, 2019, PERFORMANCE EVALUATI, DOI [10.20944/preprints201908.0019.v1, DOI 10.20944/PREPRINTS201908.0019.V1]
  • [8] Baccianella S, 2010, LREC 2010 - SEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION
  • [9] Badaro Gilbert., 2018, PROC 12 INT WORKSHOP, P236, DOI DOI 10.18653/V1/S18-1036
  • [10] Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, 10.48550/arXiv.1409.0473, DOI 10.48550/ARXIV.1409.0473]