Enhancing the Analysis of Interdisciplinary Learning Quality with GPT Models: Fine-Tuning and Knowledge-Empowered Approaches

被引:1
作者
Zhong, Tianlong [1 ]
Cai, Chang [1 ]
Zhu, Gaoxia [1 ]
Ma, Min [1 ]
机构
[1] Nanyang Technol Univ, 50 Nanyang Ave, Singapore 639798, Singapore
来源
ARTIFICIAL INTELLIGENCE IN EDUCATION: POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS, DOCTORAL CONSORTIUM AND BLUE SKY, AIED 2024 | 2024年 / 2151卷
关键词
GPT; Prompt engineering; Fine-tuning; Interdisciplinary learning; Qualitative coding;
D O I
10.1007/978-3-031-64312-5_19
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Assessing the interdisciplinary learning quality of student learning processes is significant but complex. While some research has experimented with ChatGPT for qualitative analysis of text data through crafting prompts for tasks, the in-depth consideration of task-specific knowledge, like context and rules, is still limited. The study examined whether considering such knowledge can improve ChatGPT's labeling accuracy for interdisciplinary learning quality. The data for this research consists of 252 online posts collected during class discussions. This study utilized prompt engineering, fine-tuning, and knowledge-empowered approaches to evaluate student interdisciplinary learning and compare their accuracy. The results indicated that unmodified GPT-3.5 lacks the capability for analyzing interdisciplinary learning. Fine-tuning significantly improved the models, doubling the accuracy compared to using GPT-3.5 with prompts alone. Knowledge-empowered approaches enhanced both the prompt-based and fine-tuned models, surpassing the researchers' inter-rater reliability in assessing all dimensions of student posts. This study showcased the effectiveness of combining fine-tuning and knowledge-empowered approaches with advanced language models in assessing interdisciplinary learning, indicating the potential of applying this method for qualitative analysis in educational settings.
引用
收藏
页码:157 / 165
页数:9
相关论文
共 12 条
[1]  
Brassler M, 2017, INTERDIS J PROBL-BAS, V11, DOI 10.7771/1541-5015.1686
[2]   EFAR-MMLA: An Evaluation Framework to Assess and Report Generalizability of Machine Learning Models in MMLA [J].
Chejara, Pankaj ;
Prieto, Luis P. ;
Ruiz-Calleja, Adolfo ;
Rodriguez-Triana, Maria Jesus ;
Shankar, Shashi Kant ;
Kasepalu, Reet .
SENSORS, 2021, 21 (08)
[3]  
Elliott V, 2018, QUAL REP, V23, P2850
[4]  
Fereday J., 2006, INT J QUAL METH, V5, P80, DOI DOI 10.1177/16094069060050010
[5]  
Gvili IEF, 2016, INT J ENG EDUC, V32, P123
[6]  
Liu JC, 2021, Arxiv, DOI arXiv:2101.06804
[7]   Pre-train, Prompt, and Predict: A Systematic Survey of Prompting Methods in Natural Language Processing [J].
Liu, Pengfei ;
Yuan, Weizhe ;
Fu, Jinlan ;
Jiang, Zhengbao ;
Hayashi, Hiroaki ;
Neubig, Graham .
ACM COMPUTING SURVEYS, 2023, 55 (09)
[8]  
Liu WJ, 2020, AAAI CONF ARTIF INTE, V34, P2901
[9]   Prompt Programming for Large Language Models: Beyond the Few-Shot Paradigm [J].
Reynolds, Laria ;
McDonell, Kyle .
EXTENDED ABSTRACTS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS (CHI'21), 2021,
[10]  
Shelke M. S., 2017, Int J Recent Trends Eng Res, V3, P444, DOI DOI 10.23883/IJRTER.2017.3168.0UWXM