Incorporating Pre-trained Transformer Models into TextCNN for Sentiment Analysis on Software Engineering Texts

被引:3
|
作者
Sun, Kexin [1 ]
Shi, XiaoBo [2 ]
Gao, Hui [1 ]
Kuang, Hongyu [1 ]
Ma, Xiaoxing [1 ]
Rong, Guoping [1 ]
Shao, Dong [1 ]
Zhao, Zheng [3 ]
Zhang, He [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Dalian Maritime Univ, Coll Informat Sci & Technol, Dalian, Peoples R China
[3] Dalian Maritime Univ, Coll Artificial Intelligence, Dalian, Peoples R China
来源
13TH ASIA-PACIFIC SYMPOSIUM ON INTERNETWARE, INTERNETWARE 2022 | 2022年
基金
中国国家自然科学基金;
关键词
Sentiment Analysis; Pre-trained Models; Software Mining; Nature Language Processing;
D O I
10.1145/3545258.3545273
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Software information sites (e.g., Jira, Stack Overflow) are now widely used in software development. These online platforms for collaborative development preserve a large amount of Software Engineering (SE) texts. These texts enable researchers to detect developers' attitudes toward their daily development by analyzing the sentiments expressed in the texts. Unfortunately, recent works reported that neither off-the-shelf tools nor SE-specified tools for sentiment analysis on SE texts can provide satisfying and reliable results. In this paper, we propose to incorporate pre-trained transformer models into the sentence-classification oriented deep learning framework named TextCNN to better capture the unique expression of sentiments in SE texts. Specifically, we introduce an optimized BERT model named RoBERTa as the word embedding layer of TextCNN, along with additional residual connections between RoBERTa and TextCNN for better cooperation in our training framework. An empirical evaluation based on four datasets from different software information sites shows that our training framework can achieve overall better accuracy and generalizability than the four baselines.
引用
收藏
页码:127 / 136
页数:10
相关论文
共 50 条
  • [1] Sentiment Analysis for Software Engineering: How Far Can Pre-trained Transformer Models Go?
    Zhang, Ting
    Xu, Bowen
    Thung, Ferdian
    Haryono, Stefanus Agus
    Lo, David
    Jiang, Lingxiao
    2020 IEEE INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME 2020), 2020, : 70 - 80
  • [2] Aspect Based Sentiment Analysis using French Pre-Trained Models
    Essebbar, Abderrahman
    Kane, Bamba
    Guinaudeau, Ophelie
    Chiesa, Valeria
    Quenel, Ilhem
    Chau, Stephane
    ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 1, 2021, : 519 - 525
  • [3] Explainable Pre-Trained Language Models for Sentiment Analysis in Low-Resourced Languages
    Mabokela, Koena Ronny
    Primus, Mpho
    Celik, Turgay
    BIG DATA AND COGNITIVE COMPUTING, 2024, 8 (11)
  • [4] Efficient utilization of pre-trained models: A review of sentiment analysis via prompt learning
    Bu, Kun
    Liu, Yuanchao
    Ju, Xiaolong
    KNOWLEDGE-BASED SYSTEMS, 2024, 283
  • [5] Evaluating pre-trained models for user feedback analysis in software engineering: a study on classification of app-reviews
    Hadi, Mohammad A.
    Fard, Fatemeh H. H.
    EMPIRICAL SOFTWARE ENGINEERING, 2023, 28 (04)
  • [6] Fusion Pre-trained Emoji Feature Enhancement for Sentiment Analysis
    Chen, Jie
    Yao, Zhiqiang
    Zhao, Shu
    Zhang, Yanping
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (04)
  • [7] Sentiment analysis based on improved pre-trained word embeddings
    Rezaeinia, Seyed Mahdi
    Rahmani, Rouhollah
    Ghodsi, Ali
    Veisi, Hadi
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 117 : 139 - 147
  • [8] AraXLNet: pre-trained language model for sentiment analysis of Arabic
    Alduailej, Alhanouf
    Alothaim, Abdulrahman
    JOURNAL OF BIG DATA, 2022, 9 (01)
  • [9] AraXLNet: pre-trained language model for sentiment analysis of Arabic
    Alhanouf Alduailej
    Abdulrahman Alothaim
    Journal of Big Data, 9
  • [10] A Study of Vietnamese Sentiment Classification with Ensemble Pre-trained Language Models
    Thin, Dang Van
    Hao, Duong Ngoc
    Nguyen, Ngan Luu-Thuy
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2024, 11 (01) : 137 - 165