Can Small Language Models With Retrieval-Augmented Generation Replace Large Language Models When Learning Computer Science?

被引:4
作者
Liu, Suqing [1 ]
Yu, Zezhu [1 ]
Huang, Feiran [1 ]
Bulbulia, Yousef [1 ]
Bergen, Andreas [1 ]
Liut, Michael [1 ]
机构
[1] Univ Toronto Mississauga, Mississauga, ON, Canada
来源
PROCEEDINGS OF THE 2024 CONFERENCE INNOVATION AND TECHNOLOGY IN COMPUTER SCIENCE EDUCATION, VOL 1, ITICSE 2024 | 2024年
关键词
Small Language Models; Retrieval Augmented Generation; Large Language Models; Intelligence Concentration; Conversational Agent; Personalized AI Agent; Locally Deployable AI; Intelligent Tutoring System; Intelligent Teaching Assistant; CS1; Computing Education;
D O I
10.1145/3649217.3653554
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Leveraging Large Language Models (LLMs) for personalized learning and support is becoming a promising tool in computing education. AI Assistants can help students with programming, problem-solving, converse with them to clarify course content, explain error messages to help with debugging, and much more. However, using cloud-based LLMs poses risks around data security, privacy, but also control of the overarching system. To address these concerns, we created a locally-stored Small Language Model (SLM) that leverages different Retrieval-Augmented Generation (RAG) methods to support computing students' learning. We compare one SLM (neural-chat-7b-v3 - fine-tuned version of Mistral-7B-v0.1) against two popular LLMs (gpt-3.5-turbo and gpt-4-32k) to see the viability for computing educators to use in their course(s). We use conversations from a CS1 course (N = 1, 260), providing students with an AI Assistant (using gpt-3.5-turbo) to help them learn content and support problem-solving while completing their Python programming assignment. In total, we had 269 students use the AI Assistant, with a total of 1, 988 questions asked. Using this real conversational data, we re-ran student questions using our novel SLM (neural-chat-7b-v3 testing nine different RAG methods) and gpt-4-32k, then compared those results against the original gpt-3.5-turbo responses. Our findings indicate that using an SLM with RAG can perform similarly, if not better, than LLMs. This shows that it is possible for computing educators to use SLMs (with RAG) in their course(s) as a tool for scalable learning, supporting content understanding and problem-solving needs, while employing their own policies on data privacy and security.
引用
收藏
页码:388 / 393
页数:6
相关论文
共 45 条
  • [41] Milvus: A Purpose-Built Vector Data Management System
    Wang, Jianguo
    Yi, Xiaomeng
    Guo, Rentong
    Jin, Hai
    Xu, Peng
    Li, Shengjun
    Wang, Xiangyu
    Guo, Xiangzhou
    Li, Chengming
    Xu, Xiaohai
    Yu, Kun
    Yuan, Yuxing
    Zou, Yinghao
    Long, Jiquan
    Cai, Yudong
    Li, Zhenxiang
    Zhang, Zhifeng
    Mo, Yihua
    Gu, Jun
    Jiang, Ruiyi
    Wei, Yi
    Xie, Charles
    [J]. SIGMOD '21: PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, 2021, : 2614 - 2627
  • [42] Wei JS, 2022, Arxiv, DOI [arXiv:2206.07682, DOI 10.48550/ARXIV.2206.07682]
  • [43] Yacef K, 2002, INTERNATIONAL CONFERENCE ON COMPUTERS IN EDUCATION, VOLS I AND II, PROCEEDINGS, P136, DOI 10.1109/CIE.2002.1185885
  • [44] Yu S., 2021, INTRO ARTIFICIAL INT
  • [45] Zhou Z., 2020, ICMLT 2020 P 2020 5, P18, DOI [10.1145/3409073.3409079, DOI 10.1145/3409073.3409079]