Large Language Models are Versatile Decomposers: Decomposing Evidence and Questions for Table-based Reasoning

被引:12
|
作者
Ye, Yunhu [1 ,4 ]
Hui, Binyuan [2 ]
Yang, Min [3 ]
Li, Binhua [2 ]
Huang, Fei [2 ]
Li, Yongbin [2 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] DAMO Acad, Alibaba Grp, Hangzhou, Peoples R China
[3] Chinese Acad Sci, SIAT, Shenzhen, Peoples R China
[4] Chinese Acad Sci, Shenzhen Inst Adv Technol SIAT, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023 | 2023年
关键词
Table-based reasoning; Large language models; Pre-trained language models;
D O I
10.1145/3539618.3591708
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Table-based reasoning has shown remarkable progress in a wide range of table-based tasks. It is a challenging task, which requires reasoning over both free-form natural language (NL) questions and (semi-)structured tabular data. However, previous table-based reasoning solutions usually suffer from significant performance degradation on "huge" evidence (tables). In addition, most existing methods struggle to reason over complex questions since the essential information is scattered in different places. To alleviate the above challenges, we exploit large language models (LLMs) as decomposers for effective table-based reasoning, which (i) decompose huge evidence (a huge table) into sub-evidence (a small table) to mitigate the interference of useless information for table reasoning, and (ii) decompose a complex question into simpler sub-questions for text reasoning. First, we use a powerful LLM to decompose the evidence involved in the current question into the sub-evidence that retains the relevant information and excludes the remaining irrelevant information from the "huge" evidence. Second, we propose a novel "parsing-execution-filling" strategy to decompose a complex question into simper step-by-step sub-questions by generating intermediate SQL queries as a bridge to produce numerical and logical sub-questions with a powerful LLM. Finally, we leverage the decomposed sub-evidence and sub-questions to get the final answer with a few in-context prompting examples. Extensive experiments on three benchmark datasets (TabFact, WikiTableQuestion, and FetaQA) demonstrate that our method achieves significantly better results than competitive baselines for table-based reasoning. Notably, our method outperforms human performance for the first time on the TabFact dataset. In addition to impressive overall performance, our method also has the advantage of interpretability, where the returned results are to some extent tractable with the generated sub-evidence and sub-questions. For reproducibility, we release our source code and data at: https://github.com/AlibabaResearch/DAMO-ConvAI.
引用
收藏
页码:174 / 184
页数:11
相关论文
共 50 条
  • [31] Better Results Through Ambiguity Resolution: Large Language Models that Ask Clarifying Questions
    Tix, Bernadette
    Binsted, Kim
    AUGMENTED COGNITION, PT II, AC 2024, 2024, 14695 : 72 - 87
  • [32] Performance of two large language models for data extraction in evidence synthesis
    Konet, Amanda
    Thomas, Ian
    Gartlehner, Gerald
    Kahwati, Leila
    Hilscher, Rainer
    Kugley, Shannon
    Crotty, Karen
    Viswanathan, Meera
    Chew, Robert
    RESEARCH SYNTHESIS METHODS, 2024, 15 (05) : 818 - 824
  • [33] RELAND: Integrating Large Language Models' Insights into Industrial Recommenders via a Controllable Reasoning Pool
    Tian, Changxin
    Hu, Binbin
    Gan, Chunjing
    Chen, Haoyu
    Zhang, Zhuo
    Yu, Li
    Liu, Ziqi
    Zhang, Zhiqiang
    Zhou, Jun
    Chen, Jiawei
    PROCEEDINGS OF THE EIGHTEENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2024, 2024, : 63 - 73
  • [34] Using large language models for safety-related table summarization in clinical study reports
    Landman, Rogier
    Healey, Sean P.
    Loprinzo, Vittorio
    Kochendoerfer, Ulrike
    Winnier, Angela Russell
    Henstock, Peter, V
    Lin, Wenyi
    Chen, Aqiu
    Rajendran, Arthi
    Penshanwar, Sushant
    Khan, Sheraz
    Madhavan, Subha
    JAMIA OPEN, 2024, 7 (02)
  • [35] Leveraging Non-Parametric Reasoning With Large Language Models for Enhanced Knowledge Graph Completion
    Zhang, Ying
    Shen, Yangpeng
    Xiao, Gang
    Peng, Jinghui
    IEEE ACCESS, 2024, 12 : 177012 - 177027
  • [36] Based on Medicine, The Now and Future of Large Language Models
    Su, Ziqing
    Tang, Guozhang
    Huang, Rui
    Qiao, Yang
    Zhang, Zheng
    Dai, Xingliang
    CELLULAR AND MOLECULAR BIOENGINEERING, 2024, 17 (04) : 263 - 277
  • [37] NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models
    Zhou, Gengze
    Hong, Yicong
    Wang, Zun
    Wang, Xin Eric
    Wu, Qi
    COMPUTER VISION-ECCV 2024, PT VII, 2025, 15065 : 260 - 278
  • [38] Leveraging large language models to construct feedback from medical multiple-choice Questions
    Tomova, Mihaela
    Rosello Atanet, Ivan
    Sehy, Victoria
    Sieg, Miriam
    Maerz, Maren
    Maeder, Patrick
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [39] The impact of chatbots based on large language models on second language vocabulary acquisition
    Zhang, Zhihui
    Huang, Xiaomeng
    HELIYON, 2024, 10 (03)
  • [40] An Intent-based Networks Framework based on Large Language Models
    Fuad, Ahlam
    Ahmed, Azza H.
    Riegler, Michael A.
    Cicic, Tarik
    2024 IEEE 10TH INTERNATIONAL CONFERENCE ON NETWORK SOFTWARIZATION, NETSOFT 2024, 2024, : 7 - 12