Data Augmentation Using BERT-Based Models for Aspect-Based Sentiment Analysis

被引:0
|
作者
Hollander, Bron [1 ]
Frasincar, Flavius [1 ]
van der Knaap, Finn [1 ]
机构
[1] Erasmus Univ, Burgemeester Oudlaan 50, NL-3062 PA Rotterdam, Netherlands
来源
WEB ENGINEERING, ICWE 2024 | 2024年 / 14629卷
关键词
Aspect-based sentiment classification; Data augmentation; Neural network; Pre-trained language model;
D O I
10.1007/978-3-031-62362-2_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In today's digital world, there is an overwhelming amount of opinionated data on the Web. However, effectively analyzing all available data proves to be a resource-intensive endeavor, requiring substantial time and financial investments to curate high-quality training datasets. To mitigate such problems, this paper compares data augmentation models for aspect-based sentiment analysis. Specifically, we analyze the effect of several BERT-based data augmentation methods on the performance of the state-of-the-art HAABSA++ model. We consider the following data augmentation models: EDA-adjusted (baseline), BERT, Conditional-BERT, BERTprepend, and BERTexpand. Our findings show that incorporating data augmentation techniques can significantly improve the out-of-sample accuracy of the HAABSA++ model. Specifically, our results highlight the effectiveness of BERTprepend and BERTexpand, increasing the test accuracy from 78.56% to 79.23% and from 82.62% to 84.47% for the SemEval 2015 and SemEval 2016 datasets, respectively.
引用
收藏
页码:115 / 122
页数:8
相关论文
empty
未找到相关数据