Large Language Models are Versatile Decomposers: Decomposing Evidence and Questions for Table-based Reasoning

被引:12
|
作者
Ye, Yunhu [1 ,4 ]
Hui, Binyuan [2 ]
Yang, Min [3 ]
Li, Binhua [2 ]
Huang, Fei [2 ]
Li, Yongbin [2 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] DAMO Acad, Alibaba Grp, Hangzhou, Peoples R China
[3] Chinese Acad Sci, SIAT, Shenzhen, Peoples R China
[4] Chinese Acad Sci, Shenzhen Inst Adv Technol SIAT, Shenzhen, Peoples R China
来源
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023 | 2023年
关键词
Table-based reasoning; Large language models; Pre-trained language models;
D O I
10.1145/3539618.3591708
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Table-based reasoning has shown remarkable progress in a wide range of table-based tasks. It is a challenging task, which requires reasoning over both free-form natural language (NL) questions and (semi-)structured tabular data. However, previous table-based reasoning solutions usually suffer from significant performance degradation on "huge" evidence (tables). In addition, most existing methods struggle to reason over complex questions since the essential information is scattered in different places. To alleviate the above challenges, we exploit large language models (LLMs) as decomposers for effective table-based reasoning, which (i) decompose huge evidence (a huge table) into sub-evidence (a small table) to mitigate the interference of useless information for table reasoning, and (ii) decompose a complex question into simpler sub-questions for text reasoning. First, we use a powerful LLM to decompose the evidence involved in the current question into the sub-evidence that retains the relevant information and excludes the remaining irrelevant information from the "huge" evidence. Second, we propose a novel "parsing-execution-filling" strategy to decompose a complex question into simper step-by-step sub-questions by generating intermediate SQL queries as a bridge to produce numerical and logical sub-questions with a powerful LLM. Finally, we leverage the decomposed sub-evidence and sub-questions to get the final answer with a few in-context prompting examples. Extensive experiments on three benchmark datasets (TabFact, WikiTableQuestion, and FetaQA) demonstrate that our method achieves significantly better results than competitive baselines for table-based reasoning. Notably, our method outperforms human performance for the first time on the TabFact dataset. In addition to impressive overall performance, our method also has the advantage of interpretability, where the returned results are to some extent tractable with the generated sub-evidence and sub-questions. For reproducibility, we release our source code and data at: https://github.com/AlibabaResearch/DAMO-ConvAI.
引用
收藏
页码:174 / 184
页数:11
相关论文
共 50 条
  • [41] ToM-LM: Delegating Theory of Mind Reasoning to External Symbolic Executors in Large Language Models
    Tang, Weizhi
    Belle, Vaishak
    NEURAL-SYMBOLIC LEARNING AND REASONING, PT II, NESY 2024, 2024, 14980 : 245 - 257
  • [42] Enabling controllable table-to-text generation via prompting large language models with guided planning
    Zhao, Shuo
    Sun, Xin
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [43] Selecting from Multiple Strategies Improves the Foreseeable Reasoning of Tool-Augmented Large Language Models
    Wu, Yongchao
    Henriksson, Aron
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT III, ECML PKDD 2024, 2024, 14943 : 197 - 212
  • [44] Applications of large language models in cancer care: current evidence and future perspectives
    Iannantuono, Giovanni Maria
    Bracken-Clarke, Dara
    Floudas, Charalampos S.
    Roselli, Mario
    Gulley, James L.
    Karzai, Fatima
    FRONTIERS IN ONCOLOGY, 2023, 13
  • [45] Usage-based constructionist approaches and large language models
    Goldberg, Adele E.
    CONSTRUCTIONS AND FRAMES, 2024, 16 (02) : 220 - 254
  • [46] Ethical framework for AI education based on large language models
    Yuyang Yan
    Hui Liu
    Education and Information Technologies, 2025, 30 (8) : 10891 - 10909
  • [47] Embedding Search for Quranic Texts based on Large Language Models
    Alqarni, Mohammed
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2024, 21 (02) : 243 - 256
  • [48] Shopfloor layout generation method based on large language models
    Hu, Yi
    Sun, Yicheng
    Wen, Xiaojian
    Wang, Sen
    Bao, Jinsong
    INTERNATIONAL JOURNAL OF COMPUTER INTEGRATED MANUFACTURING, 2025,
  • [49] Code Clone Detection Techniques Based on Large Language Models
    Almatrafi, Afnan A.
    Eassa, Fathy A.
    Sharaf, Sanaa A.
    IEEE ACCESS, 2025, 13 : 46136 - 46146
  • [50] Accuracy of large language models in answering ophthalmology board-style questions: A meta-analysis
    Wu, Jo-Hsuan
    Nishida, Takashi
    Liu, T. Y. Alvin
    ASIA-PACIFIC JOURNAL OF OPHTHALMOLOGY, 2024, 13 (05):