Rationality of Thought Improves Reasoning in Large Language Models

被引:0
作者
Gou, Tian [1 ,2 ]
Zhang, Boyao [1 ,2 ]
Sun, Zhenglie [1 ,2 ]
Wang, Jing [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Yangang [1 ,2 ]
Wang, Jue [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2024 | 2024年 / 14887卷
基金
北京市自然科学基金; 国家重点研发计划;
关键词
Large Language Models (LLMs); Zero-Shot Reasoning; Cognitive foundations of knowledge; Rationality of Thought (RoT); Cognitive Psychology; Cognitive Bias Dataset; HEURISTICS; FALLACY;
D O I
10.1007/978-981-97-5501-1_26
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the capabilities of large language models (LLMs) have been progressively advanced, their competence in addressing intricate reasoning tasks remains inadequate, primarily due to their insufficient cognitive capabilities. To explore the cognitive proficiency of models like GPT-4, we turn to methodologies from cognitive psychology: cognitive abilities reflect rational thinking skills, and cognitive bias tasks are often used to assess rational thinking levels. In this paper, we develop a cognitive bias dataset to measure the rational thinking and cognitive levels of LLMs. Our observations indicate that GPT-4, akin to humans, exhibits limitations in its rational thinking ability. We propose a new method, "Rationality of Thought" (RoT), to prompt LLMs into a rational thinking process during task execution. This method significantly improves the accuracy of GPT-4 on the cognitive bias task by 18.7%. Cognitive capacity is also essential for tackling complex issues, therefore, we implement RoT across various reasoning tasks. Using only a zero-shot setting, RoT outperforms inference enhancement techniques such as CoT using zero-shot, such as SVAMP(+1.8),AQUA-RAT (+6.0), ARC-c (+4.1),ARCe(+3.9) in multiple arithmetic and common sense reasoning tasks. Our empirical evaluation shows that RoT helps LLMs elevate their cognitive capabilities through rational thinking, thereby becoming more adept at navigating complex reasoning tasks.
引用
收藏
页码:343 / 358
页数:16
相关论文
共 50 条
[41]   Responsible Use of Large Language Models: An Analogy with the Oxford Tutorial System [J].
Lissack, Michael ;
Meagher, Brenden .
SHE JI-THE JOURNAL OF DESIGN ECONOMICS AND INNOVATION, 2024, 10 (04) :389-413
[42]   Agile Methodology for the Standardization of Engineering Requirements Using Large Language Models [J].
Ray, Archana Tikayat ;
Cole, Bjorn F. ;
Fischer, Olivia Pinon J. ;
Bhat, Anirudh Prabhakara ;
White, Ryan T. ;
Mavris, Dimitri N. .
SYSTEMS, 2023, 11 (07)
[43]   AI am a rheumatologist: a practical primer to large language models for rheumatologists [J].
Venerito, Vincenzo ;
Bilgin, Emre ;
Iannone, Florenzo ;
Kiraz, Sedat .
RHEUMATOLOGY, 2023, 62 (10) :3256-3260
[44]   The role of large language models in medical image processing: a narrative review [J].
Tian, Dianzhe ;
Jiang, Shitao ;
Zhang, Lei ;
Lu, Xin ;
Xu, Yiyao .
QUANTITATIVE IMAGING IN MEDICINE AND SURGERY, 2024, 14 (01) :1108-1121
[45]   Mechanistic interpretability of large language models with applications to the financial services industry [J].
Golgoon, Ashkan ;
Filom, Khashayar ;
Kannan, Arjun Ravi .
5TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2024, 2024, :660-668
[46]   The Impact of Large Language Models on Programming Education and Student Learning Outcomes [J].
Jost, Gregor ;
Taneski, Viktor ;
Karakatic, Saso .
APPLIED SCIENCES-BASEL, 2024, 14 (10)
[47]   Towards automated phenotype definition extraction using large language models [J].
Ramya Tekumalla ;
Juan M. Banda .
Genomics & Informatics, 22 (1)
[48]   Prompt Wrangling: On Replication and Generalization in Large Language Models for PCG Levels [J].
Karkaj, Arash Moradi ;
Nelson, Mark J. ;
Koutis, Ioannis ;
Hoover, Amy K. .
PROCEEDINGS OF THE 19TH INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF DIGITAL GAMES, FDG 2024, 2024,
[49]   Large Language Models and the Enhancement of Human Cognition: Some Theoretical Insights [J].
Dirzyte, Aiste .
FILOSOFIJA-SOCIOLOGIJA, 2025, 36 (01) :14-22
[50]   Improving requirements completeness: automated assistance through large language models [J].
Luitel, Dipeeka ;
Hassani, Shabnam ;
Sabetzadeh, Mehrdad .
REQUIREMENTS ENGINEERING, 2024, 29 (01) :73-95