The Problem of AI Hallucination and How to Solve It

被引:0
|
作者
Jancarik, Antonin [1 ]
Dusek, Ondrej [2 ]
机构
[1] Charles Univ Prague, Fac Educ, Prague, Czech Republic
[2] Charles Univ Prague, Fac Math & Phys, Prague, Czech Republic
来源
PROCEEDINGS OF THE 23RD EUROPEAN CONFERENCE ON E-LEARNING, ECEL 2024 | 2024年 / 23/1卷
关键词
Chatbots; AI; Mathematics education; Hallucination; ARTIFICIAL-INTELLIGENCE;
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
AI in education is a topic that has been researched for the last 70 years. However, the last two years have seen very significant changes. These changes relate to the introduction of OpenAI's ChatGPT chatbot in November 2022. The GPT (Generative Pre-trained Transformer) language model has dramatically influenced how the public approaches artificial intelligence. For many, generative language models have become synonymous with AI and have come uncritically viewed as a universal source of answers to most questions. However, it soon became apparent that even generative language models had their limits. Among the main problems that emerged was hallucination (providing answers containing false or misleading information), which is expected in all language models. The main problem of hallucination is that this information is difficult to distinguish from other information, and AI language models are very persuasive in presenting it. The risks of this phenomenon are much more substantial when using language modules to support learning, where the learner cannot distinguish correct information from incorrect information. The proposed paper focuses on the area of AI hallucination in mathematics education. It will first show how AI chatbots hallucinate in mathematics and then present one possible solution to counter this hallucination. The presented solution was created for the AI chatbot Edu-AI and designed to tutor students in mathematics. Usually, the problem is approached so that the system verifies the correctness of the output offered by the chatbot. Within the Edu-AI, checking responses is not implemented, but checking inputs is. If an input containing a factual query is recorded, it is redirected, and the answer is traced to authorised knowledge sources and study materials. If a relevant answer cannot be traced in these sources, a redirect to a natural person who will address the question is offered. In addition to describing the technical solution, the article includes concrete examples of how the system works. This solution has been developed for the educational domain but applies to all domains where users must be provided with relevant information.
引用
收藏
页码:122 / 128
页数:7
相关论文
共 50 条
  • [1] Hallucination in AI-generated financial literature reviews: evaluating bibliographic accuracy
    Erdem, Orhan
    Hassett, Kristi
    Egriboyun, Feyzullah
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2025,
  • [2] How do People Perceive the Role of AI in Human-AI Collaboration to Solve Everyday Tasks?
    Papachristos, Eleftherios
    Johansen, Patrick Skov
    Jacobsen, Rune Moberg
    Bysted, Lukas Bjorn Leer
    Skov, Mikael B.
    PROCEEDINGS OF 1ST INTERNATIONAL CONFERENCE OF THE GREECE ACM SIGCHI CHAPTER, CHIGREECE 2021, 2021,
  • [3] No amount of "AI" in content moderation will solve filtering's prior-restraint problem
    Llanso, Emma J.
    BIG DATA & SOCIETY, 2020, 7 (01):
  • [4] Using AI to solve business problems in scholarly publishing
    Upshall, Michael
    INSIGHTS-THE UKSG JOURNAL, 2019, 32
  • [5] Deep Automation Bias: How to Tackle a Wicked Problem of AI?
    Strauss, Stefan
    BIG DATA AND COGNITIVE COMPUTING, 2021, 5 (02)
  • [6] How Can the Clinical Aptitude of AI Assistants Be Assayed?
    Thirunavukarasu, Arun James
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2023, 25
  • [7] Embers of autoregression show how large language models are shaped by the problem they are trained to solve
    McCoy, R. Thomas
    Yao, Shunyu
    Friedman, Dan
    Hardy, Mathew D.
    Griffiths, Thomas L.
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2024, 121 (41)
  • [8] Busier emergency departments: same triage systems-can AI solve the "next-to-be-seen" problem?
    Holodinsky, Jessalyn K.
    Sabir, Lisa
    Mcrae, Andrew D.
    Mason, Suzanne M.
    CANADIAN JOURNAL OF EMERGENCY MEDICINE, 2025, 27 (01) : 5 - 6
  • [9] The AI-augmented crowd: How human crowdvoters adopt AI (or not)
    Freisinger, Elena
    Unfried, Matthias
    Schneider, Sabrina
    JOURNAL OF PRODUCT INNOVATION MANAGEMENT, 2024, 41 (04) : 865 - 889
  • [10] Coming to Terms with the Black Box Problem: How to Justify AI Systems in Health Care
    Felder, Ryan Marshall
    HASTINGS CENTER REPORT, 2021, 51 (04) : 38 - 45