ChatGPT improves creative problem-solving performance in university students: An experimental study

被引:108
作者
Urban, Marek [1 ]
Dechterenko, Filip [1 ,2 ]
Lukavsky, Jiri [1 ,2 ]
Hrabalova, Veronika [2 ]
Svacha, Filip [3 ]
Brom, Cyril [4 ]
Urban, Kamila [5 ]
机构
[1] Czech Acad Sci, Inst Psychol, Hybernska 8, Prague 11000 1, Czech Republic
[2] Charles Univ Prague, Fac Arts, Nam J Palacha 1-2, Prague 11638, Czech Republic
[3] Charles Univ Prague, Fac Humanities, Patkova 2137-5, Prague 18200, Czech Republic
[4] Charles Univ Prague, Fac Math & Phys, V Holesovickach 747-2, Prague, Czech Republic
[5] Slovak Acad Sci, Inst Res Social Commun, Dubravska Cesta 9, Bratislava 84511, Slovakia
关键词
Generative AI; ChatGPT; Creativity; Metacognitive monitoring; Metacognitive experiences; Ill-defined problem -solving task; SELF-EFFICACY; ANTECEDENTS; MOTIVATION;
D O I
10.1016/j.compedu.2024.105031
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
University students often employ generative artificial intelligence tools such as ChatGPT in resolution of ill-defined problem-solving tasks. However, the experimental evidence about effects of ChatGPT on complex problem-solving performance is still missing. In this preregistered experiment, the impact of ChatGPT on performance in a complex creative problem-solving task was investigated in 77 university students solving a task with ChatGPT in comparison to 68 students solving a task without it. ChatGPT use significantly improved self-efficacy for task resolution (d = 0.65) and enhanced the quality (d = 0.69), elaboration (d = 0.61), and originality (d = 0.55) of solutions. Moreover, participants with ChatGPT assistance perceived task as easier (d = 0.56) and requiring less mental effort (d = 0.58). However, use of ChatGPT did not make task resolution more interesting (d = 0.08), and the impact of ChatGPT on metacognitive monitoring accuracy was unclear. Although there were no significant differences in absolute accuracy between students solving the task with and without the assistance of ChatGPT, the absence of correlation between self-evaluation judgments and performance suggests that participants struggled to calibrate their self-evaluations when using ChatGPT. Notably, the perceived usefulness of ChatGPT appeared to inform self-evaluation judgments, resulting in higher inaccuracy. The implications for hybrid human-AI regulation (HHAIR) theory are discussed. To regulate effectively, students using AI tools should focus on valid metacognitive cues instead of the perceived ease of ChatGPT-assisted problem-solving.
引用
收藏
页数:15
相关论文
共 82 条
[1]  
Ackerman R, 2019, PSIHOLOGIJSKE TEME, V28, P1, DOI [10.31820/pt.28.1.1, 10.31820/pt.28.1.1, DOI 10.31820/PT.28.1.1]
[2]  
Alexander S., 2022, Astral Codex Ten
[3]  
Brown TB, 2020, Arxiv, DOI [arXiv:2005.14165, DOI 10.48550/ARXIV.2005.14165]
[4]   The Relation Between Student's Effort and Monitoring Judgments During Learning: A Meta-analysis [J].
Baars, Martine ;
Wijnia, Lisette ;
de Bruin, Anique ;
Paas, Fred .
EDUCATIONAL PSYCHOLOGY REVIEW, 2020, 32 (04) :979-1002
[5]  
Bandura A., 1999, SELF SOCIAL PSYCHOL, P285, DOI DOI 10.2307/2073538
[6]  
Bang Y, 2023, Arxiv, DOI [arXiv:2302.04023, 10.48550/arXiv.2302.04023]
[7]  
Beghetto R.A., 2017, The Creative Self: Effect of Beliefs, Self-Efficacy, Mindset, and Identity, P3, DOI DOI 10.1016/B978-0-12-809790-8.00001-7
[8]  
Bezirhan U, 2023, Computers and Education Artificial Intelligence, V5, P100161, DOI [10.1016/j.caeai.2023.100161, 10.1016/j.caeai.2023.100161, DOI 10.1016/J.CAEAI.2023.100161]
[9]  
Brown G.T. L., 2013, SAGE handbook of research on classroom assessment, P367, DOI DOI 10.4135/9781452218649.N21
[10]  
Dang H, 2022, Arxiv, DOI [arXiv:2209.01390, 10.48550/arxiv.2209.01390]