Do Large Language Models Show Human-like Biases? Exploring Confidence-Competence Gap in AI

被引:2
作者
Singh, Aniket Kumar [1 ]
Lamichhane, Bishal [2 ]
Devkota, Suman [3 ]
Dhakal, Uttam [3 ]
Dhakal, Chandra [4 ]
机构
[1] Youngstown State Univ, Dept Comp Sci & Informat Syst, Youngstown, OH 44555 USA
[2] Univ Nevada, Dept Math & Stat, Reno, NV 89557 USA
[3] Youngstown State Univ, Dept Elect & Comp Engn, Youngstown, OH 44555 USA
[4] Univ Georgia, Dept Agr & Appl Econ, Athens, GA 30602 USA
关键词
Large Language Models; Dunning-Kruger effects; chat-GPT; BARD; Claude; LLaMA; cognitive biases; artificial intelligence; AI ethics; Natural Language Processing; confidence assessment;
D O I
10.3390/info15020092
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This study investigates self-assessment tendencies in Large Language Models (LLMs), examining if patterns resemble human cognitive biases like the Dunning-Kruger effect. LLMs, including GPT, BARD, Claude, and LLaMA, are evaluated using confidence scores on reasoning tasks. The models provide self-assessed confidence levels before and after responding to different questions. The results show cases where high confidence does not correlate with correctness, suggesting overconfidence. Conversely, low confidence despite accurate responses indicates potential underestimation. The confidence scores vary across problem categories and difficulties, reducing confidence for complex queries. GPT-4 displays consistent confidence, while LLaMA and Claude demonstrate more variations. Some of these patterns resemble the Dunning-Kruger effect, where incompetence leads to inflated self-evaluations. While not conclusively evident, these observations parallel this phenomenon and provide a foundation to further explore the alignment of competence and confidence in LLMs. As LLMs continue to expand their societal roles, further research into their self-assessment mechanisms is warranted to fully understand their capabilities and limitations.
引用
收藏
页数:20
相关论文
共 24 条