Generative Chain-of-Thought for Zero-Shot Cognitive Reasoning

被引:0
作者
Liu, Liang [1 ]
Zhang, Dong [1 ]
Zhu, Suyang [2 ]
Li, Shoushan [1 ]
机构
[1] Soochow Univ, Sch Comp Sci & Technol, Suzhou 215006, Peoples R China
[2] Suzhou City Univ, Comp Sci & Artificial Intelligence Coll, Suzhou 215104, Peoples R China
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT V | 2024年 / 15020卷
关键词
Zero-shot Cognitive Reasoning; Generative; Chain of Thought; Large Language Models;
D O I
10.1007/978-3-031-72344-5_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Cognitive reasoning holds a significant place within the field of Natural Language Processing (NLP). Yet, the exploration of zeroshot scenarios, which align more closely with real-life situations than supervised scenarios, has been relatively limited. While a few studies have employed Large Language Models (LLMs) to tackle zero-shot cognitive reasoning tasks, they still grapple with two key challenges: 1) Traditional approaches rely on the chain-of-thought (CoT) mechanism, wherein LLMs are provided with a "'think step by step" prompt. However, this zero-shot learning approach may not effectively leverage multiple similar demonstrations and may be susceptible to errors. 2) Previous CoT methods have predominantly focused on intricate mathematical reasoning tasks, overlooking the fact that conventional NLP tasks can also be reframed as cognitive and reasoning processes, such as sentiment analysis and question answering tasks. Consequently, LLMs can be harnessed for zero-shot cognitive reasoning problems in NLP. To address these issues, we introduce a generative CoT approach for performing zero-shot cognitive reasoning tasks. Our experimental results clearly demonstrate that our approach can outperform the existing state-of-the-art methods across three categories of tasks: sentiment analysis, question answering, and mathematical reasoning.
引用
收藏
页码:324 / 339
页数:16
相关论文
共 36 条
  • [1] Brown TB, 2020, ADV NEUR IN, V33
  • [2] Chowdhery A, 2023, J MACH LEARN RES, V24
  • [3] Chung HW, 2022, Arxiv, DOI [arXiv:2210.11416, DOI 10.48550/ARXIV.2210.11416]
  • [4] Deng PF, 2022, Arxiv, DOI [arXiv:2209.02276, DOI 10.48550/ARXIV.2209.02276]
  • [5] In two minds: dual-process accounts of reasoning
    Evans, JST
    [J]. TRENDS IN COGNITIVE SCIENCES, 2003, 7 (10) : 454 - 459
  • [6] Fei H, 2023, 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, P1171
  • [7] Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies
    Geva, Mor
    Khashabi, Daniel
    Segal, Elad
    Khot, Tushar
    Roth, Dan
    Berant, Jonathan
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2021, 9 : 346 - 361
  • [8] Hoffmann J, 2022, Arxiv, DOI arXiv:2203.15556
  • [9] Hosseini Mohammad Javad, 2014, P 2014 C EMP METH NA, P523, DOI DOI 10.3115/V1/D14-1058
  • [10] Jiang AQ, 2023, Arxiv, DOI arXiv:2310.06825