Eliciting metaknowledge in Large Language Models

被引:0
作者
Longo, Carmelo Fabio [1 ]
Mongiovi, Misael [1 ,2 ]
Bulla, Luana [1 ,2 ]
Lieto, Antonio [3 ,4 ]
机构
[1] CNR, Inst Sci & Technol Cognit, Rome, Italy
[2] Univ Catania, Dept Math & Comp Sci, Catania, Italy
[3] Univ Salerno, Cognit Interact & Intelligent Technol Lab DISPC, Fisciano, Italy
[4] CNR, ICAR, Rome, Italy
来源
COGNITIVE SYSTEMS RESEARCH | 2025年 / 91卷
关键词
Artificial Intelligence; Natural Language Processing; Question answering; Large Language Models; Meta-knowledge;
D O I
10.1016/j.cogsys.2025.101352
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The introduction of Large Language Models (LLMs) able to exhibit a number of linguistic and extra-linguistic capabilities has represented, in the last years, one of the main frontiers in Artificial Intelligence (AI) research. Researcher from various disciplines debate about whether or not, among the capabilities of LLMs, there is the one of using knowledge about knowledge-usually considered one of the antechambers of meta-cognition in cognitive agents-about a particular task in order to improve or self-correct previous errors. In this work we propose a novel fine-tuning approach for LLMs, named EXAR, based on a multi-stage process leveraging past predictions from an early version of the same, and aimed at injecting metacognitive features for the task of Question-Answering. The conducted experiments on LLAMA-2-7B-CHAT showed promising improvements on the quality of the outcomes, due to the fact that the LLM acquired the ability to detect its own wrong predictions forcing itself to repeat submissions, thorough a prompt designed to fix inadmissible predictions, whenever detected. Such detection is achieved by enquiring the same LLM acting as meta-validator, through another prompt specifically designed for such purpose.
引用
收藏
页数:8
相关论文
共 24 条
[1]  
Bakhtin A, 2020, Arxiv, DOI arXiv:2004.10188
[2]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[3]  
Diao SZ, 2024, Arxiv, DOI [arXiv:2302.12246, 10.48550/arXiv.2302.12246]
[4]  
Driess D, 2023, Arxiv, DOI [arXiv:2303.03378, 10.48550/arXiv.2303.03378, DOI 10.48550/ARXIV.2303.03378]
[5]  
Gao YF, 2024, Arxiv, DOI [arXiv:2312.10997, DOI 10.48550/ARXIV.2312.10997]
[6]  
He P., 2021, ARXIV
[7]  
Holtzman A, 2018, Arxiv, DOI arXiv:1805.06087
[8]  
Hu EJ, 2021, Arxiv, DOI [arXiv:2106.09685, DOI 10.48550/ARXIV.2106.09685]
[9]   View-independent representation with frame interpolation method for skeleton-based human action recognition [J].
Jiang, Yingguo ;
Xu, Jun ;
Zhang, Tong .
INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2020, 11 (12) :2625-2636
[10]  
Kim G, 2023, Arxiv, DOI [arXiv:2303.17491, 10.48550/arXiv.2303.17491, DOI 10.48550/ARXIV.2303.17491]