Chain-of-Thought Reasoning in Tabular Language Models

被引:0
作者
Zheng, Mingyu [1 ,2 ,3 ]
Hao, Yang [3 ]
Jiang, Wenbin [3 ]
Lin, Zheng [1 ,2 ]
Lyu, Yajuan [3 ]
She, Qiaoqiao [3 ]
Wang, Weiping [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Baidu Inc, Beijing, Peoples R China
来源
FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023) | 2023年
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tabular mathematical reasoning task requires models to perform multi-step operations including information look-up and numerical calculations, based on heterogeneous data from tables and questions. Existing solutions tend to extend chain-of-thought (CoT) reasoning into powerful large language models (LLMs) to promote multi-hop mathematical reasoning. However, it can be extremely difficult to apply such LLMbased approaches under scenarios of privatization deployment or limited resources. To address this problem, we revisit small-scale tabular language models (TaLMs) and extend chainof-thought reasoning into TaLMs for the first time. Specifically, we propose a novel framework, TaCo, which coordinates two TaLMs responsible for CoT generation and answer inference, respectively. Besides, our framework can be combined with an external calculator to enhance accurate numerical calculations. On the TABMWP dataset, TaCo outperforms the state-of-the-art ChatGPT by 9.55% (82.60%.92.15% in accuracy) with much less parameters (0.8B).(1)
引用
收藏
页码:11006 / 11019
页数:14
相关论文
共 37 条
  • [1] On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?
    Bender, Emily M.
    Gebru, Timnit
    McMillan-Major, Angelina
    Shmitchell, Shmargaret
    [J]. PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 610 - 623
  • [2] Brown TB, 2020, ADV NEUR IN, V33
  • [3] Chen M., 2021, arXiv
  • [4] Chen Wenhu, 2022, Program of thoughts prompting: Disentangling computation from reasoning for numerical reasoning tasks
  • [5] Chen ZY, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P3697
  • [6] Chowdhery A., 2022, PaLM: Scaling language modeling with pathways
  • [7] Cohen William W., 2021, MATE: multi-view attention for table transformer efficiency
  • [8] Dong HY, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P5426
  • [9] Fu Yao, 2023, Complexity-based prompting for multi-step reasoning
  • [10] Gong H., 2020, P 28 INT C COMP LING, P1978