Recall Mechanism and Multi-Head Attention for Numerical Reasoning

被引:0
作者
Lai, Linjia [1 ,2 ,3 ]
Tan, Tien-Ping [2 ]
Zeng, Bocan [2 ]
机构
[1] Putian Univ, New Engn Ind Coll, Putian 351100, Peoples R China
[2] Univ Sains Malaysia, Sch Comp Sci, Minden 11800, Malaysia
[3] Putian Univ, Putian Elect Informat Ind Technol Res Inst, Putian 351100, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2025年 / 15卷 / 07期
关键词
artificial intelligence; financial report analysis; natural language processing; numerical reasoning; multi-head attention; recall mechanism;
D O I
10.3390/app15073528
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Numerical reasoning is a challenging question-answering task in artificial intelligence (AI) that requires both reading comprehension and numerical computation capabilities. Although recent approaches have made significant progress in reasoning, two critical issues remain: (1) information tends to be gradually lost as the network deepens due to the complexity of deep learning models, and (2) the performance of multi-step reasoning is suboptimal, leaving room for improvement in accuracy. To address these issues, we propose a model with a recall mechanism and multi-head attention for numerical reasoning (RMMANR). The recall mechanism prevents the embedding information of questions and passages from being forgotten as the model deepens, while the multi-head attention mechanism analyzes possible solutions for numerical reasoning. RMMANR consists of two main components: an encoder and a decoder. The encoder leverages RoBERTa to encode the question and passage into contextual embeddings. The decoder, which consists of four modules (RM Module, Selector, MA Module, and Program Solver), generates inference steps and answers based on contextual information from the encoder. We implement our model using PyTorch and evaluate it on the FINQA dataset, a benchmark for numerical reasoning in the financial domain. Experimental results demonstrate that RMMANR outperforms several baseline models, achieving superior accuracy.
引用
收藏
页数:24
相关论文
共 40 条
  • [1] BERT Models for Arabic Text Classification: A Systematic Review
    Alammary, Ali Saleh
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (11):
  • [2] Callanan E., 2024, P 8 FIN TECHN NAT LA, P23
  • [3] Chen D, 2014, P C EMP METH NAT LAN, P740, DOI 10.3115/v1/d14-1082
  • [4] Chen KL, 2023, Arxiv, DOI arXiv:2009.07448
  • [5] Multi-head attention based candidate segment selection in QA over hybrid data
    Chen, Qian
    Gao, Xiaoying
    Guo, Xin
    Wang, Suge
    [J]. INTELLIGENT DATA ANALYSIS, 2023, 27 (06) : 1839 - 1852
  • [6] Chen WH, 2023, Arxiv, DOI arXiv:2211.12588
  • [7] Chen X., 2019, P INT C LEARN REPR N
  • [8] Chen ZY, 2022, Arxiv, DOI arXiv:2109.00122
  • [9] A Review of Sentiment, Semantic and Event-Extraction-Based Approaches in Stock Forecasting
    Cheng, Wai Khuen
    Bea, Khean Thye
    Leow, Steven Mun Hong
    Chan, Jireh Yi-Le
    Hong, Zeng-Wei
    Chen, Yen-Lin
    [J]. MATHEMATICS, 2022, 10 (14)
  • [10] Chung JY, 2014, Arxiv, DOI [arXiv:1412.3555, DOI 10.48550/ARXIV.1412.3555]