Fine-Tuning a Large Language Model with Reinforcement Learning for Educational Question Generation

被引:0
|
作者
Lamsiyah, Salima [1 ]
El Mahdaouy, Abdelkader [2 ]
Nourbakhsh, Aria [1 ]
Schommer, Christoph [1 ]
机构
[1] Univ Luxembourg, Fac Sci Technol & Med, Dept Comp Sci, Esch Sur Alzette, Luxembourg
[2] Mohammed VI Polytech Univ, Coll Comp, Ben Guerir, Morocco
来源
ARTIFICIAL INTELLIGENCE IN EDUCATION, PT I, AIED 2024 | 2024年 / 14829卷
关键词
Educational Question Generation; Large Language Model; Google FLAN-T5; Reinforcement Learning; Self-Critical Sequence Training;
D O I
10.1007/978-3-031-64302-6_30
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Educational Natural Language Generation (EduQG) aims to automatically generate educational questions from textual content, which is crucial for the expansion of online education. Prior research in EduQG has predominantly relied on cross-entropy loss for training, which can lead to issues such as exposure bias and inconsistencies between training and testing metrics. To mitigate this issue, we propose a reinforcement learning (RL) based large language model (LLM) for educational question generation. In particular, we fine-tune the Google FLAN-T5 model using a mixed objective function that combines cross-entropy and RL losses to ensure the generation of questions that are syntactically and semantically accurate. The experimental results on the SciQ question generation dataset show that the proposed method is competitive with current state-of-the-art systems in terms of predictive performance and linguistic quality.
引用
收藏
页码:424 / 438
页数:15
相关论文
共 50 条
  • [31] Fine-tuning natural language imperatives
    Kaufmann, Magdalena
    JOURNAL OF LOGIC AND COMPUTATION, 2019, 29 (03) : 321 - 348
  • [32] On Surgical Fine-tuning for Language Encoders
    Lodha, Abhilasha
    Belapurkar, Gayatri
    Chalkapurkar, Saloni
    Tao, Yuanming
    Ghosh, Reshmi
    Basu, Samyadeep
    Petrov, Dmitrii
    Srinivasan, Soundararajan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 3105 - 3113
  • [33] CRaSh: Clustering, Removing, and Sharing Enhance Fine-tuning without Full Large Language Model
    Zhang, Kaiyan
    Ding, Ning
    Qi, Biqing
    Zhu, Xuekai
    Long, Xinwei
    Zhou, Bowen
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9612 - 9637
  • [34] Fine-Tuning Large Language Models for Private Document Retrieval: A Tutorial
    Sommers, Frank
    Kongthon, Alisa
    Kongyoung, Sarawoot
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 1319 - 1320
  • [35] Large language models in Radiology: The importance of fine-tuning and the fable of the luthier
    Martin-Noguerol, Teodoro
    Lopez-Ubeda, Pilar
    Luna, Antonio
    EUROPEAN JOURNAL OF RADIOLOGY, 2024, 178
  • [36] Enhancing AAV Viability Prediction: A Generalizable Fine-Tuning Framework with Large Language Model xTrimoPGLM
    Yang, Qirong
    Zou, Diming
    Guo, Yucheng
    Xu, Chenrui
    Marsic, Damien
    Cai, Zhongshan
    Liu, Yawen
    Xu, Ziyao
    Qu, Vicky
    Garces, Fernando
    Greisen, Per
    Ji, Qingzhou
    Song, Le
    MOLECULAR THERAPY, 2024, 32 (04) : 693 - 694
  • [37] Distributed Inference and Fine-tuning of Large Language Models Over The Internet
    Borzunov, Alexander
    Ryabinin, Max
    Chumachenko, Artem
    Baranchuk, Dmitry
    Dettmers, Tim
    Belkada, Younes
    Samygin, Pavel
    Raffel, Colin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] Fine-Tuning Large Enterprise Language Models via Ontological Reasoning
    Baldazzi, Teodoro
    Bellomarini, Luigi
    Ceri, Stefano
    Colombo, Andrea
    Gentili, Andrea
    Sallinger, Emanuel
    RULES AND REASONING, RULEML+RR 2023, 2023, 14244 : 86 - 94
  • [39] Repeatability of Fine-Tuning Large Language Models Illustrated Using QLoRA
    Alahmari, Saeed S.
    Hall, Lawrence O.
    Mouton, Peter R.
    Goldgof, Dmitry B.
    IEEE ACCESS, 2024, 12 : 153221 - 153231
  • [40] Fine-tuning large language models for rare disease concept normalization
    Wang, Andy
    Liu, Cong
    Yang, Jingye
    Weng, Chunhua
    JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 2076 - 2083