Math-LLMs: AI Cyberinfrastructure with Pre-trained Transformers for Math Education

被引:3
|
作者
Zhang, Fan [1 ]
Li, Chenglu [2 ]
Henkel, Owen [3 ]
Xing, Wanli [1 ]
Baral, Sami [4 ]
Heffernan, Neil [4 ]
Li, Hai [1 ]
机构
[1] Univ Florida, Gainesville, FL 32611 USA
[2] Univ Utah, Salt Lake City, UT USA
[3] Rising Acad Network, Freetown, Sierra Leone
[4] Worcester Polytech Inst, Worcester, MA USA
关键词
LLMs; Math education; Pre-train; MATHEMATICS;
D O I
10.1007/s40593-024-00416-y
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In recent years, the pre-training of Large Language Models (LLMs) in the educational domain has garnered significant attention. However, a discernible gap exists in the application of these models to mathematics education. This study aims to bridge this gap by pre-training LLMs on authentic K-12 mathematical dialogue datasets. Our research is structured around three primary research questions (RQs) that investigate the impact of fine-tuning data size and pre-training in downstream Natural Language Processing (NLP) tasks, and the efficacy of LLMs in text generation tasks within the mathematical context. Our findings indicate that data size plays a pivotal role in the performance of LLMs in downstream NLP tasks, with larger datasets yielding more consistent and improved results. Furthermore, pre-trained models consistently outperformed their non-pre-trained counterparts, emphasizing the importance of leveraging prior knowledge in LLMs. In the realm of text generation, we found that our model can not only enhance mathematical understanding and performance on downstream math tasks but also generate more engaging and human-like language.
引用
收藏
页数:24
相关论文
共 4 条
  • [1] EAPT: An encrypted traffic classification model via adversarial pre-trained transformers
    Zhan, Mingming
    Yang, Jin
    Jia, Dongqing
    Fu, Geyuan
    COMPUTER NETWORKS, 2025, 257
  • [2] Parameter-Efficient Multi-classification Software Defect Detection Method Based on Pre-trained LLMs
    Wang, Xuanye
    Lu, Lu
    Yang, Zhanyu
    Tian, Qingyan
    Lin, Haisha
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2024, 17 (01)
  • [3] Impacts of interacting with an AI chatbot on preservice teachers' responsive teaching skills in math education
    Lee, Dabae
    Son, Taekwon
    Yeo, Sheunghyun
    JOURNAL OF COMPUTER ASSISTED LEARNING, 2025, 41 (01)
  • [4] MNIST-Fraction: Enhancing Math Education with AI-Driven Fraction Detection and Analysis
    Ahadian, Pegah
    Feng, Yunhe
    Kosko, Karl
    Ferdig, Richard
    Guan, Qiang
    PROCEEDINGS OF THE 2024 ACM SOUTHEAST CONFERENCE, ACMSE 2024, 2024, : 284 - 290