Incorporating Pre-trained Transformer Models into TextCNN for Sentiment Analysis on Software Engineering Texts

被引:3
|
作者
Sun, Kexin [1 ]
Shi, XiaoBo [2 ]
Gao, Hui [1 ]
Kuang, Hongyu [1 ]
Ma, Xiaoxing [1 ]
Rong, Guoping [1 ]
Shao, Dong [1 ]
Zhao, Zheng [3 ]
Zhang, He [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing, Peoples R China
[2] Dalian Maritime Univ, Coll Informat Sci & Technol, Dalian, Peoples R China
[3] Dalian Maritime Univ, Coll Artificial Intelligence, Dalian, Peoples R China
来源
13TH ASIA-PACIFIC SYMPOSIUM ON INTERNETWARE, INTERNETWARE 2022 | 2022年
基金
中国国家自然科学基金;
关键词
Sentiment Analysis; Pre-trained Models; Software Mining; Nature Language Processing;
D O I
10.1145/3545258.3545273
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Software information sites (e.g., Jira, Stack Overflow) are now widely used in software development. These online platforms for collaborative development preserve a large amount of Software Engineering (SE) texts. These texts enable researchers to detect developers' attitudes toward their daily development by analyzing the sentiments expressed in the texts. Unfortunately, recent works reported that neither off-the-shelf tools nor SE-specified tools for sentiment analysis on SE texts can provide satisfying and reliable results. In this paper, we propose to incorporate pre-trained transformer models into the sentence-classification oriented deep learning framework named TextCNN to better capture the unique expression of sentiments in SE texts. Specifically, we introduce an optimized BERT model named RoBERTa as the word embedding layer of TextCNN, along with additional residual connections between RoBERTa and TextCNN for better cooperation in our training framework. An empirical evaluation based on four datasets from different software information sites shows that our training framework can achieve overall better accuracy and generalizability than the four baselines.
引用
收藏
页码:127 / 136
页数:10
相关论文
共 50 条
  • [21] A Comparative Study of Different Pre-trained Language Models for Sentiment Analysis of Human-Computer Negotiation Dialogue
    Dong, Jing
    Luo, Xudong
    Zhu, Junlin
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2024, 2024, 14887 : 301 - 317
  • [22] The Biases of Pre-Trained Language Models: An Empirical Study on Prompt-Based Sentiment Analysis and Emotion Detection
    Mao, Rui
    Liu, Qian
    He, Kai
    Li, Wei
    Cambria, Erik
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 1743 - 1753
  • [23] Incorporating Transformer Models for Sentiment Analysis and News Classification in Khmer
    Rifat, Md Rifatul Islam
    Al Imran, Abdullah
    COMPUTATIONAL DATA AND SOCIAL NETWORKS, CSONET 2021, 2021, 13116 : 106 - 117
  • [24] Sentiment analysis of COP9-related tweets: a comparative study of pre-trained models and traditional techniques
    Elmitwalli, Sherif
    Mehegan, John
    FRONTIERS IN BIG DATA, 2024, 7
  • [25] SsciBERT: a pre-trained language model for social science texts
    Si Shen
    Jiangfeng Liu
    Litao Lin
    Ying Huang
    Lin Zhang
    Chang Liu
    Yutong Feng
    Dongbo Wang
    Scientometrics, 2023, 128 : 1241 - 1263
  • [26] SsciBERT: a pre-trained language model for social science texts
    Shen, Si
    Liu, Jiangfeng
    Lin, Litao
    Huang, Ying
    Zhang, Lin
    Liu, Chang
    Feng, Yutong
    Wang, Dongbo
    SCIENTOMETRICS, 2023, 128 (02) : 1241 - 1263
  • [27] Pre-trained models: Past, present and future
    Han, Xu
    Zhang, Zhengyan
    Ding, Ning
    Gu, Yuxian
    Liu, Xiao
    Huo, Yuqi
    Qiu, Jiezhong
    Yao, Yuan
    Zhang, Ao
    Zhang, Liang
    Han, Wentao
    Huang, Minlie
    Jin, Qin
    Lan, Yanyan
    Liu, Yang
    Liu, Zhiyuan
    Lu, Zhiwu
    Qiu, Xipeng
    Song, Ruihua
    Tang, Jie
    Wen, Ji-Rong
    Yuan, Jinhui
    Zhao, Wayne Xin
    Zhu, Jun
    AI OPEN, 2021, 2 : 225 - 250
  • [28] Natural Attack for Pre-trained Models of Code
    Yang, Zhou
    Shi, Jieke
    He, Junda
    Lo, David
    2022 ACM/IEEE 44TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING (ICSE 2022), 2022, : 1482 - 1493
  • [29] HinPLMs: Pre-trained Language Models for Hindi
    Huang, Xixuan
    Lin, Nankai
    Li, Kexin
    Wang, Lianxi
    Gan, Suifu
    2021 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP), 2021, : 241 - 246
  • [30] An analysis of pre-trained stable diffusion models through a semantic lens
    Bonechi, Simone
    Andreini, Paolo
    Corradini, Barbara Toniella
    Scarselli, Franco
    NEUROCOMPUTING, 2025, 614