Using GPT-4 to guide causal machine learning

被引:0
作者
Constantinou, Anthony C. [1 ]
Kitson, Neville K. [1 ]
Zanga, Alessio [1 ,2 ,3 ]
机构
[1] Queen Mary Univ London, Sch Elect Engn & Comp Sci, Bayesian AI Res Lab, Machine Intelligence & Decis Syst MInDS Res Grp, London E1 4NS, England
[2] Univ Milano Bicocca, Dept Informat Syst & Commun, Models & Algorithms Data & Text Min Lab MADLab, Milan, Italy
[3] F Hoffmann La Roche Ltd, Data Sci & Adv Analyt, Basel, Switzerland
关键词
Bayesian networks; Causal discovery; ChatGPT; Directed acyclic graphs; Knowledge graphs; LLMs; Structure learning;
D O I
10.1016/j.eswa.2024.126120
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since its introduction to the public, ChatGPT has had an unprecedented impact. While some experts praised AI advancements and highlighted their potential risks, others have been critical about the accuracy and usefulness of Large Language Models (LLMs). In this paper, we are interested in the ability of LLMs to identify causal relationships. We focus on the well-established GPT-4 (Turbo) and evaluate its performance under the most restrictive conditions, by isolating its ability to infer causal relationships based solely on the variable labels without being given any other context by humans, demonstrating the minimum level of effectiveness one can expect when it is provided with label-only information. We show that questionnaire participants judge the GPT-4 graphs as the most accurate in the evaluated categories, closely followed by knowledge graphs constructed by domain experts, with causal Machine Learning (ML) far behind. We use these results to highlight the important limitation of causal ML, which often produces causal graphs that violate common sense, affecting trust in them. However, we show that pairing GPT-4 with causal ML overcomes this limitation, resulting in graphical structures learnt from real data that align more closely with those identified by domain experts, compared to structures learnt by causal ML alone. Overall, our findings suggest that despite GPT-4 not being explicitly designed to reason causally, it can still be a valuable tool for causal representation, as it improves the causal discovery process of causal ML algorithms that are designed to do just that.
引用
收藏
页数:15
相关论文
共 29 条
  • [1] Antonucci A., 2023, P 38 AAAI C ART INT
  • [2] Bubeck S, 2023, Arxiv, DOI [arXiv:2303.12712, DOI 10.48550/ARXIV.2303.12712]
  • [3] Cohrs K., 2004, AAAI 2024 WORKSH LLM
  • [4] Constantinou A., 2019, The Bayesys user manual
  • [5] Constantinou A.C., 2020, The Bayesys data and Bayesian network repository
  • [6] Open problems in causal structure learning: A case study of COVID-19 in the UK
    Constantinou, Anthony
    Kitson, Neville K.
    Liu, Yang
    Chobtham, Kiattikun
    Amirkhizi, Arian Hashemzadeh
    Nanavati, Praharsh A.
    Mbuvha, Rendani
    Petrungaro, Bruno
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 234
  • [7] The impact of prior knowledge on causal structure learning
    Constantinou, Anthony C.
    Guo, Zhigao
    Kitson, Neville K.
    [J]. KNOWLEDGE AND INFORMATION SYSTEMS, 2023, 65 (08) : 3385 - 3434
  • [8] The importance of temporal information in Bayesian network structure learning
    Constantinou, Anthony C.
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2021, 164 (164)
  • [9] Jin Z., 2024, P 12 INT C LEARN REP, P2024
  • [10] Jiralerspong T, 2024, Arxiv, DOI arXiv:2402.01207