DeCoT: Debiasing Chain-of-Thought for Knowledge-Intensive Tasks in Large Language Models via Causal Intervention

被引:0
|
作者
Wu, Junda [1 ]
Yu, Tong [2 ]
Chen, Xiang [2 ]
Wang, Haoliang [2 ]
Rossi, Ryan A. [2 ]
Kim, Sungchul [2 ]
Rao, Anup [2 ]
McAuley, Julian [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA 92093 USA
[2] Adobe Res, San Jose, CA USA
来源
PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS | 2024年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) often require task-relevant knowledge to augment their internal knowledge through prompts. However, simply injecting external knowledge into prompts does not guarantee that LLMs can identify and use relevant information in the prompts to conduct chain-of-thought reasoning, especially when the LLM's internal knowledge is derived from biased information on the pretraining data. In this paper, we propose a novel causal view to formally explain the internal knowledge bias of LLMs via a Structural Causal Model (SCM). We review the chain-of-thought (CoT) prompting from a causal perspective, and discover that the biased information from pretrained models can impair LLMs' reasoning abilities. When the CoT reasoning paths are misled by irrelevant information from prompts and are logically incorrect, simply editing factual information is insufficient to reach the correct answer. To estimate the confounding effect on CoT reasoning in LLMs, we use external knowledge as an instrumental variable. We further introduce CoT as a mediator to conduct front-door adjustment and generate logically correct CoTs where the spurious correlation between LLMs' pretrained knowledge and task queries is reduced. With extensive experiments, we validate that our approach enables more accurate CoT reasoning and enhances LLM generation on knowledge-intensive tasks.
引用
收藏
页码:14073 / 14087
页数:15
相关论文
共 22 条
  • [21] Improving intermediate reasoning in zero-shot chain-of-thought for large language models with filter supervisor-self correction
    Sun, Jun
    Pan, Yiteng
    Yan, Xiaohu
    NEUROCOMPUTING, 2025, 620
  • [22] The COT COLLECTION: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning
    Kim, Seungone
    Jool, Se June
    Kim, Doyoung
    Jang, Joel
    Ye, Seonghyeon
    Shin, Jamin
    Seo, Minjoon
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12685 - 12708