Enhancing Financial Sentiment Analysis Ability of Language Model via Targeted Numerical Change-Related Masking

被引:4
作者
Jung, Hui Do [1 ]
Jang, Beakcheol [1 ]
机构
[1] Yonsei Univ, Grad Sch Informat, Seoul 03722, South Korea
基金
新加坡国家研究基金会;
关键词
Task analysis; Numerical models; Predictive models; Analytical models; Sentiment analysis; Data models; Training; Financial sentiment analysis; language model; numerical understanding; post-training; PREDICTION;
D O I
10.1109/ACCESS.2024.3385855
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Sentiment analysis is a critical task that is highly beneficial to various financial tasks such as stock-price prediction, corporate credit rating, economic report analysis, and investment decision support. Researchers have used various methods to train pretraining language models (PLMs) for these tasks. Although most PLMs have achieved excellent performance, they can be further improved. In this study, we propose a new framework to strengthen numerical understanding, in particular for the FinBERT(Financial Bidirectional Encoder Representations from Transformers) model released in 2019, thus improving model performance in the task of sentiment analysis on financial news sentences. This method selects sentences containing numerical words from financial news articles, preferentially masks the words, and post-train the PLM. To evaluate the proposed methodology quantitatively, we apply the same post-training to different financial language models and compare the performance before and after the application using Financial Phrasebank, which is a representative benchmark dataset used in financial sentiment analysis. The experimental results show that the best performance is achieved when 50,000 sentences are used to post-train FinBERT, thus confirming the advantage of the proposed methodology for downstream tasks and highlighting the importance of using the correct amount of data. Additionally, we show that applying the proposed method to different language models improves the performance, particularly in low-resource environments with less training data. The findings of this study suggest that the PLM can improve aspects that it does not understand well, and that thd PLM performance can be improved by post-training it with task- and domain-appropriate datasets, in not only finance but also in other domains.
引用
收藏
页码:50809 / 50820
页数:12
相关论文
共 48 条
[1]  
Akita A., 2016, IEEEACIS 15 INT C CO, P1
[2]  
Araci D, 2019, Arxiv, DOI [arXiv:1908.10063, DOI 10.48550/ARXIV.1908.10063]
[3]   Sentiment analysis of the Spanish financial stability Report* [J].
Bernal, Angel Ivan Moreno ;
Pedraz, Carlos Gonzalez .
INTERNATIONAL REVIEW OF ECONOMICS & FINANCE, 2024, 89 :913-939
[4]  
Cortis K., 2017, P 11 INT WORKSH SEM, P519
[5]   Evaluating and understanding text-based stock price prediction models [J].
de Fortuny, Enric Junque ;
De Smedt, Tom ;
Martens, David ;
Daelemans, Walter .
INFORMATION PROCESSING & MANAGEMENT, 2014, 50 (02) :426-441
[6]  
Devlin J, 2019, Arxiv, DOI arXiv:1810.04805
[7]   A Neural Network Model for Semi-supervised Review Aspect Identification [J].
Ding, Ying ;
Yu, Changlong ;
Jiang, Jing .
ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2017, PT II, 2017, 10235 :668-680
[8]  
Du CN, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P4019
[9]  
Goldmann L., 2022, Ph.D. thesis, DOI [10.7488/era/2023, DOI 10.7488/ERA/2023]
[10]  
Hasselgren B, 2023, FUTURE INTERNET, V15, DOI [10.3390/fi15010005, 10.22069/FPPJ.2023.21395.1763]