Chain-of-Thought Reasoning in Tabular Language Models

被引:0
作者
Zheng, Mingyu [1 ,2 ,3 ]
Hao, Yang [3 ]
Jiang, Wenbin [3 ]
Lin, Zheng [1 ,2 ]
Lyu, Yajuan [3 ]
She, Qiaoqiao [3 ]
Wang, Weiping [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Baidu Inc, Beijing, Peoples R China
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tabular mathematical reasoning task requires models to perform multi-step operations including information look-up and numerical calculations, based on heterogeneous data from tables and questions. Existing solutions tend to extend chain-of-thought (CoT) reasoning into powerful large language models (LLMs) to promote multi-hop mathematical reasoning. However, it can be extremely difficult to apply such LLMbased approaches under scenarios of privatization deployment or limited resources. To address this problem, we revisit small-scale tabular language models (TaLMs) and extend chainof-thought reasoning into TaLMs for the first time. Specifically, we propose a novel framework, TaCo, which coordinates two TaLMs responsible for CoT generation and answer inference, respectively. Besides, our framework can be combined with an external calculator to enhance accurate numerical calculations. On the TABMWP dataset, TaCo outperforms the state-of-the-art ChatGPT by 9.55% (82.60%.92.15% in accuracy) with much less parameters (0.8B).(1)
引用
收藏
页码:11006 / 11019
页数:14
相关论文
共 37 条
  • [31] Wei JS, 2022, ADV NEUR IN
  • [32] Xing XY, 2021, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, P2273
  • [33] Yao S., 2023, TREE THOUGHTS DELIBE
  • [34] Yao Yao, 2023, Beyond chain-of-thought, effective graph-of-thought reasoning in large language models
  • [35] Zhang Zhuosheng, 2022, AUTOMATIC CHAIN THOU
  • [36] Zhao Ruochen, 2023, Verify-and-edit: A knowledge-enhanced chain-of-thought framework
  • [37] Zhu Fengbin, 2021, arXiv