Incorporating Pre-trained Transformer Models into TextCNN for Sentiment Analysis on Software Engineering Texts

被引:3
|
作者
Sun, Kexin [1 ]
Shi, XiaoBo [2 ]
Gao, Hui [1 ]
Kuang, Hongyu [1 ]
Ma, Xiaoxing [1 ]
Rong, Guoping [1 ]
Shao, Dong [1 ]
Zhao, Zheng [3 ]
Zhang, He [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Dalian Maritime Univ, Coll Informat Sci & Technol, Dalian, Peoples R China
[3] Dalian Maritime Univ, Coll Artificial Intelligence, Dalian, Peoples R China
来源
13TH ASIA-PACIFIC SYMPOSIUM ON INTERNETWARE, INTERNETWARE 2022 | 2022年
基金
中国国家自然科学基金;
关键词
Sentiment Analysis; Pre-trained Models; Software Mining; Nature Language Processing;
D O I
10.1145/3545258.3545273
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Software information sites (e.g., Jira, Stack Overflow) are now widely used in software development. These online platforms for collaborative development preserve a large amount of Software Engineering (SE) texts. These texts enable researchers to detect developers' attitudes toward their daily development by analyzing the sentiments expressed in the texts. Unfortunately, recent works reported that neither off-the-shelf tools nor SE-specified tools for sentiment analysis on SE texts can provide satisfying and reliable results. In this paper, we propose to incorporate pre-trained transformer models into the sentence-classification oriented deep learning framework named TextCNN to better capture the unique expression of sentiments in SE texts. Specifically, we introduce an optimized BERT model named RoBERTa as the word embedding layer of TextCNN, along with additional residual connections between RoBERTa and TextCNN for better cooperation in our training framework. An empirical evaluation based on four datasets from different software information sites shows that our training framework can achieve overall better accuracy and generalizability than the four baselines.
引用
收藏
页码:127 / 136
页数:10
相关论文
共 50 条
  • [41] Graph-aware pre-trained language model for political sentiment analysis in Filipino social media
    Aquino, Jean Aristide
    Liew, Di Jie
    Chang, Yung-Chun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 146
  • [42] BARTpho: Pre-trained Sequence-to-Sequence Models for Vietnamese
    Nguyen Luong Tran
    Duong Minh Le
    Dat Quoc Nguyen
    INTERSPEECH 2022, 2022, : 1751 - 1755
  • [43] Performance Evaluation of Pre-trained Models in Sarcasm Detection Task
    Wang, Haiyang
    Song, Xin
    Zhou, Bin
    Wang, Ye
    Gao, Liqun
    Jia, Yan
    WEB INFORMATION SYSTEMS ENGINEERING - WISE 2021, PT II, 2021, 13081 : 67 - 75
  • [44] GAIT analysis based on GENDER detection using pre-trained models and tune parameters
    Vora C.
    Katkar V.
    Lunagaria M.
    Discover Artificial Intelligence, 2024, 4 (01):
  • [45] Hippocampus segmentation and classification for dementia analysis using pre-trained neural network models
    Priyanka, Ahana
    Ganesan, Kavitha
    BIOMEDICAL ENGINEERING-BIOMEDIZINISCHE TECHNIK, 2021, 66 (06): : 581 - 592
  • [46] Revisiting Sentiment Analysis for Software Engineering in the Era of Large Language Models
    Zhang, Ting
    Irsan, Ivana clairine
    Thung, Ferdian
    Lo, David
    ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2025, 34 (03)
  • [47] Research on Sentiment Analysis in Software Engineering
    Chen Z.-P.
    Yao H.-H.
    Cao Y.-B.
    Liu X.-Z.
    Mei H.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (05): : 2218 - 2230
  • [48] Multi-Label Conditional Generation From Pre-Trained Models
    Proszewska, Magdalena
    Wolczyk, Maciej
    Zieba, Maciej
    Wielopolski, Patryk
    Maziarka, Lukasz
    Smieja, Marek
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (09) : 6185 - 6198
  • [49] Efficient Integrated Features Based on Pre-trained Models for Speaker Verification
    Li, Yishuang
    Guan, Wenhao
    Huang, Hukai
    Miao, Shiyu
    Su, Qi
    Li, Lin
    Hong, Qingyang
    INTERSPEECH 2024, 2024, : 2140 - 2144
  • [50] Automatic Detection of Liver Cancer Using Hybrid Pre-Trained Models
    Othman, Esam
    Mahmoud, Muhammad
    Dhahri, Habib
    Abdulkader, Hatem
    Mahmood, Awais
    Ibrahim, Mina
    SENSORS, 2022, 22 (14)