Targeted training for numerical reasoning with large language models

被引:0
作者
Li, Xiao [1 ]
Liu, Sichen [1 ]
Zhu, Yin [1 ]
Cheng, Gong [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
基金
中国国家自然科学基金;
关键词
Data refining; Numerical reasoning; Large language model; Knowledge distillation;
D O I
10.1007/s10115-024-02216-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
After recent gains achieved by large language models (LLMs) on numerical reasoning tasks, it has become of interest to have LLMs teach small models to improve on numerical reasoning. Instructing LLMs to generate Chains of Thought to fine-tune small models is an established approach. However, small models are passive in this line of work and may not be able to exploit the provided training data. In this paper, we propose a novel targeted training strategy to match LLM's assistance with small models' capacities. The small model will proactively request LLM's assistance when it sifts out confusing training data. Then, LLM refines such data by successively revising reasoning steps and reducing question complexity before feeding the small model. Experiments show that this targeted training approach remarkably improves the performance of small models on a range of numerical reasoning datasets by 12-25%, making small models even competitive with some LLMs.
引用
收藏
页码:197 / 221
页数:25
相关论文
共 68 条
[1]   Deep learning-based question answering: a survey [J].
Abdel-Nabi, Heba ;
Awajan, Arafat ;
Ali, Mostafa Z. .
KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 65 (04) :1399-1485
[2]  
Anthropic, 2023, Anthropic: Claude
[3]  
Bang Yejin, 2023, arXiv
[4]  
Bi, 2024, ARXIV
[5]  
Bucilua C., 2006, P ACM SIGKDD, V12, P535
[6]  
Chen S., 2023, ARXIV
[7]   KnowPrompt: Knowledge-aware Prompt-tuning with Synergistic Optimization for Relation Extraction [J].
Chen, Xiang ;
Zhang, Ningyu ;
Xie, Xin ;
Deng, Shumin ;
Yao, Yunzhi ;
Tan, Chuanqi ;
Huang, Fei ;
Si, Luo ;
Chen, Huajun .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2778-2788
[8]  
Chung H., 2022, arXiv
[9]  
Clark P, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P3882
[10]  
Cobbe Karl, 2021, ARXIV