The Promises and Pitfalls of Large Language Models as Feedback Providers: A Study of Prompt Engineering and the Quality of AI-Driven Feedback

被引:0
作者
Jacobsen, Lucas Jasper [1 ]
Weber, Kira Elena [1 ]
机构
[1] Univ Hamburg, Fac Educ, D-20146 Hamburg, Germany
关键词
AI; feedback; prompt engineering; teacher education; ACHIEVEMENT; ENVIRONMENTS; METAANALYSIS; IMPACT;
D O I
10.3390/ai6020035
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Background/Objectives: Artificial intelligence (AI) is transforming higher education (HE), reshaping teaching, learning, and feedback processes. Feedback generated by large language models (LLMs) has shown potential for enhancing student learning outcomes. However, few empirical studies have directly compared the quality of LLM feedback with feedback from novices and experts. This study investigates (1) the types of prompts needed to ensure high-quality LLM feedback in teacher education and (2) how feedback from novices, experts, and LLMs compares in terms of quality. Methods: To address these questions, we developed a theory-driven manual to evaluate prompt quality and designed three prompts of varying quality. Feedback generated by ChatGPT-4 was assessed alongside feedback from novices and experts, who were provided with the highest-quality prompt. Results: Our findings reveal that only the best prompt consistently produced high-quality feedback. Additionally, LLM feedback outperformed novice feedback and, in the categories explanation, questions, and specificity, even surpassed expert feedback in quality while being generated more quickly. Conclusions: These results suggest that LLMs, when guided by well-crafted prompts, can serve as high-quality and efficient alternatives to expert feedback. The findings underscore the importance of prompt quality and emphasize the need for prompt design guidelines to maximize the potential of LLMs in teacher education.
引用
收藏
页数:17
相关论文
共 6 条
  • [1] The Impact of Prompt Engineering and a Generative AI-Driven Tool on Autonomous Learning: A Case Study
    Mzwri, Kovan
    Turcsanyi-Szabo, Marta
    EDUCATION SCIENCES, 2025, 15 (02):
  • [2] A Study on Performance Improvement of Prompt Engineering for Generative AI with a Large Language Model
    Park, Daeseung
    An, Gi-taek
    Kamyod, Chayapol
    Kim, Cheong Ghil
    JOURNAL OF WEB ENGINEERING, 2023, 22 (08): : 1187 - 1206
  • [3] Prompt Engineering or Fine-Tuning? A Case Study on Phishing Detection with Large Language Models
    Trad, Fouad
    Chehab, Ali
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (01): : 367 - 384
  • [4] Implementing Artificial Intelligence in Physiotherapy Education: A Case Study on the Use of Large Language Models (LLM) to Enhance Feedback
    Villagran, Ignacio
    Hernandez, Rocio
    Schuit, Gregory
    Neyem, Andres
    Fuentes-Cimma, Javiera
    Miranda, Constanza
    Hilliger, Isabel
    Duran, Valentina
    Escalona, Gabriel
    Varas, Julian
    IEEE TRANSACTIONS ON LEARNING TECHNOLOGIES, 2024, 17 : 2079 - 2090
  • [5] Prompt Engineering Approach Study for Supervised Fine-Tuned (SFT) Large Language Models (LLMs) in Spacecraft Fault Diagnosis
    Xia, Qing
    Zhao, Haotian
    Liu, Ming
    2024 3RD CONFERENCE ON FULLY ACTUATED SYSTEM THEORY AND APPLICATIONS, FASTA 2024, 2024, : 819 - 824
  • [6] How Self-Regulated Learning Is Affected by Feedback Based on Large Language Models: Data-Driven Sustainable Development in Computer Programming Learning
    Sun, Di
    Xu, Pengfei
    Zhang, Jing
    Liu, Ruqi
    Zhang, Jun
    ELECTRONICS, 2025, 14 (01):