Continual Attention Modeling for Successive Sentiment Analysis in Low-resource Scenarios

被引:0
作者
Zhang, Han [1 ]
Wang, Jing-Jing [1 ]
Luo, Jia-Min [1 ]
Zhou, Guo-Dong [1 ]
机构
[1] School of Computer Science and Technology, Soochow University, Suzhou
来源
Ruan Jian Xue Bao/Journal of Software | 2024年 / 35卷 / 12期
关键词
Adapter; attention mechanism; continual learning; low-resource scenario; sentiment analysis;
D O I
10.13328/j.cnki.jos.007057
中图分类号
学科分类号
摘要
Currently, sentiment analysis research is generally based on big data-driven models, which heavily rely on expensive annotation and computational costs. Therefore, research on sentiment analysis in low-resource scenarios is particularly urgent. However, existing research on sentiment analysis in low-resource scenarios mainly focuses on a single task, making it difficult for models to acquire external task knowledge. Therefore, this study constructs successive sentiment analysis in low-resource scenarios, aiming to allow models to learn multiple sentiment analysis tasks over time by continual learning methods. This can make full use of data from different tasks and learn sentiment information from different tasks, thus alleviating the problem of insufficient training data for a single task. There are two core problems with successive sentiment analysis in low-resource scenarios. One is preserving sentiment information for a single task, and the other is fusing sentiment information between different tasks. To solve these two problems, this study proposes continual attention modeling for successive sentiment analysis in low-resource scenarios. Sentiment masked Adapter (SMA) is first constructed, which is used to generate hard attention emotion masks for different tasks. This can preserve sentiment information for different tasks and mitigate catastrophic forgetting. Secondly, dynamic sentiment attention (DSA) is proposed, which dynamically fuses features extracted by different Adapters based on the current time step and task similarity. This can fuse sentiment information between different tasks. Experimental results on multiple datasets show that the proposed approach significantly outperforms the state-of-the-art benchmark approaches. Additionally, experimental analysis indicates that the proposed approach has the best sentiment information retention ability and sentiment information fusion ability compared to other benchmark approaches while maintaining high operational efficiency. © 2024 Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:5470 / 5486
页数:16
相关论文
共 43 条
  • [1] Hedderich MA, Lange L, Adel H, Strotgen J, Klakow D., A survey on recent approaches for natural language processing in low-resource scenarios, Proc. of the 2021 Conf. of North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 2545-2568, (2021)
  • [2] Lohar P, Xie GD, Bendechache M, Brennan R, Celeste E, Trestian R, Tal I., Irish attitudes toward COVID tracker APP & privacy: Sentiment analysis on Twitter and survey data, Proc. of the 16th Int’l Conf. on Availability, Reliability and Security, (2021)
  • [3] McCloskey M, Cohen NJ., Catastrophic interference in connectionist networks: The sequential learning problem, Psychology of Learning and Motivation, 24, pp. 109-165, (1989)
  • [4] French RM., Catastrophic forgetting in connectionist networks, Trends in Cognitive Sciences, 3, 4, pp. 128-135, (1999)
  • [5] Garrette D, Baldridge J., Learning A part-of-speech tagger from two hours of annotation, Proc. of the 2013 Conf. of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 138-147, (2013)
  • [6] Yang Z, Wu W, Yang J, Xu C, Li ZJ., Low-resource response generation with template prior, Proc. of the 2019 Conf. on Empirical Methods in Natural Language Processing and the 9th Int’l Joint Conf. on Natural Language Processing, pp. 1886-1897, (2019)
  • [7] Ke ZX, Liu B, Ma NZ, Xu H, Shu L., Achieving forgetting prevention and knowledge transfer in continual learning, Proc. of the 35th Conf. on Neural Information Processing Systems, pp. 22443-22456, (2021)
  • [8] Wei J, Zou K., EDA: Easy data augmentation techniques for boosting performance on text classification tasks, Proc. of the 2019 Conf. on Empirical Methods in Natural Language Processing and the 9th Int’l Joint Conf. on Natural Language Processing, pp. 6382-6388, (2019)
  • [9] Raiman J, Miller J., Globally normalized reader, Proc. of the 2017 Conf. on Empirical Methods in Natural Language Processing, pp. 1059-1069, (2017)
  • [10] Xie QZ, Dai ZH, Hovy E, Luong MT, Le QV., Unsupervised data augmentation for consistency training, Proc. of the 34th Int’l Conf. on Neural Information Processing Systems, (2020)