Rationality of Thought Improves Reasoning in Large Language Models

被引:0
作者
Gou, Tian [1 ,2 ]
Zhang, Boyao [1 ,2 ]
Sun, Zhenglie [1 ,2 ]
Wang, Jing [1 ,2 ]
Liu, Fang [1 ,2 ]
Wang, Yangang [1 ,2 ]
Wang, Jue [1 ,2 ]
机构
[1] Chinese Acad Sci, Comp Network Informat Ctr, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Beijing, Peoples R China
来源
KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2024 | 2024年 / 14887卷
基金
北京市自然科学基金; 国家重点研发计划;
关键词
Large Language Models (LLMs); Zero-Shot Reasoning; Cognitive foundations of knowledge; Rationality of Thought (RoT); Cognitive Psychology; Cognitive Bias Dataset; HEURISTICS; FALLACY;
D O I
10.1007/978-981-97-5501-1_26
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While the capabilities of large language models (LLMs) have been progressively advanced, their competence in addressing intricate reasoning tasks remains inadequate, primarily due to their insufficient cognitive capabilities. To explore the cognitive proficiency of models like GPT-4, we turn to methodologies from cognitive psychology: cognitive abilities reflect rational thinking skills, and cognitive bias tasks are often used to assess rational thinking levels. In this paper, we develop a cognitive bias dataset to measure the rational thinking and cognitive levels of LLMs. Our observations indicate that GPT-4, akin to humans, exhibits limitations in its rational thinking ability. We propose a new method, "Rationality of Thought" (RoT), to prompt LLMs into a rational thinking process during task execution. This method significantly improves the accuracy of GPT-4 on the cognitive bias task by 18.7%. Cognitive capacity is also essential for tackling complex issues, therefore, we implement RoT across various reasoning tasks. Using only a zero-shot setting, RoT outperforms inference enhancement techniques such as CoT using zero-shot, such as SVAMP(+1.8),AQUA-RAT (+6.0), ARC-c (+4.1),ARCe(+3.9) in multiple arithmetic and common sense reasoning tasks. Our empirical evaluation shows that RoT helps LLMs elevate their cognitive capabilities through rational thinking, thereby becoming more adept at navigating complex reasoning tasks.
引用
收藏
页码:343 / 358
页数:16
相关论文
共 50 条
  • [31] When Software Security Meets Large Language Models: A Survey
    Zhu, Xiaogang
    Zhou, Wei
    Han, Qing-Long
    Ma, Wanlun
    Wen, Sheng
    Xiang, Yang
    [J]. IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2025, 12 (02) : 317 - 334
  • [32] The rise of large language models: challenges for Critical Discourse Studies
    Gillings, Mathew
    Kohn, Tobias
    Mautner, Gerlinde
    [J]. CRITICAL DISCOURSE STUDIES, 2024,
  • [33] Evaluating the Efficacy of Large Language Models in Identifying Phishing Attempts
    Patel, Het
    Reiman, Umair
    Iqbal, Farkhund
    [J]. 2024 16TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTION, HSI 2024, 2024,
  • [34] Human bias in AI models? Anchoring effects and mitigation strategies in large language models
    Nguyen, Jeremy K.
    [J]. JOURNAL OF BEHAVIORAL AND EXPERIMENTAL FINANCE, 2024, 43
  • [35] Language models and psychological sciences
    Sartori, Giuseppe
    Orru, Graziella
    [J]. FRONTIERS IN PSYCHOLOGY, 2023, 14
  • [36] Establishing priorities for implementation of large language models in pathology and laboratory medicine
    Arvisais-Anhalt, Simone
    Gonias, Steven L.
    Murray, Sara G.
    [J]. ACADEMIC PATHOLOGY, 2024, 11 (01):
  • [37] Improving requirements completeness: automated assistance through large language models
    Dipeeka Luitel
    Shabnam Hassani
    Mehrdad Sabetzadeh
    [J]. Requirements Engineering, 2024, 29 : 73 - 95
  • [38] Leveraging Large Language Models for Generating Personalized Care Recommendations in Dementia
    Hu, Hsiang-Wei
    Lin, Yu-chun
    Chia, Chang-Hung
    Chuang, Ethan
    Yang, Cheng Ru
    [J]. 2024 IEEE INTERNATIONAL WORKSHOP ON ELECTROMAGNETICS: APPLICATIONS AND STUDENT INNOVATION COMPETITION, IWEM 2024, 2024,
  • [39] Using Large Language Models to Translate Machine Results to Human Results
    Niraula, Trishna
    Stubblefield, Jonathan
    [J]. 14TH ACM CONFERENCE ON BIOINFORMATICS, COMPUTATIONAL BIOLOGY, AND HEALTH INFORMATICS, BCB 2023, 2023,
  • [40] Incorporating Molecular Knowledge in Large Language Models via Multimodal Modeling
    Yang, Zekun
    Lv, Kun
    Shu, Jian
    Li, Zheng
    Xiao, Ping
    [J]. IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025,