Making meta-learning solve cross-prompt automatic essay scoring

被引:0
作者
Wang, Jiong [1 ]
Zhang, Qing [2 ]
Liu, Jie [2 ,3 ]
Wang, Xiaoyi [3 ]
Xu, Mingying [2 ]
Yang, Liguang [1 ]
Zhou, Jianshe [3 ]
机构
[1] Capital Normal Univ, Coll Informat Engn, Beijing 100048, Peoples R China
[2] North China Univ Technol, Sch Informat Sci, Beijing 100144, Peoples R China
[3] Capital Normal Univ, China Language Intelligence Res Ctr, Beijing 100048, Peoples R China
基金
中国国家自然科学基金;
关键词
Automated essay scoring; Domain generalization; Meta-learning;
D O I
10.1016/j.eswa.2025.126710
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The challenge of cross-prompt automatic essay scoring (AES) task is to perform well on new unseen prompt essays when the scoring model has only been trained on seen prompt essays. Existing cross-prompt AES methods focus on obtaining essay representations with prompt-invariance between seen and unseen prompts to grade the unseen prompt essays. In this way, the prompt distribution that model learned are often unconscious, which means that they may often result in negative distributional shifts due to the lack of information from target prompts. In order to maximize the model's distribution shift towards the target prompt direction, we propose optimizing the model's training process to enable conscious prompt generalization. Specifically, we propose a novel meta-learning framework under prompt generalization setting. In our method, a meta- learner selection mechanism is proposed, which is designed to directly optimize the task scheduling strategy based on the status of the meta-learner. By introducing information from target prompts during the proposed optimization process, and leveraging this information to select meta-learning states most conducive to optimize the model towards the target direction, we achieve guiding the model to generalize towards the target prompt distribution. Besides, to enhance the diversity of meta-learning training tasks and further improve the model's generalization ability, a data augmentation strategy based on large language models is designed. We conducted experiments on the ASAP dataset, the experimental results show that the proposed approach achieves a leading average result compared than other cross-prompt AES methods.
引用
收藏
页数:11
相关论文
共 38 条
  • [1] Alikaniotis D, 2016, PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, P715
  • [2] [Anonymous], 2009, Advances in Neural Information Processing Systems 22
  • [3] Meta-Learning with Task-Adaptive Loss Function for Few-Shot Learning
    Baik, Sungyong
    Choi, Janghoon
    Kim, Heewon
    Cho, Dohee
    Min, Jaesik
    Lee, Kyoung Mu
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 9445 - 9454
  • [4] Domain-Adaptive Neural Automated Essay Scoring
    Cao, Yue
    Jin, Hanqi
    Wan, Xiaojun
    Yu, Zhiwei
    [J]. PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1011 - 1020
  • [5] Chen Y., 2024, P 2024 JOINT INT C C, P12775
  • [6] Chen Y, 2023, PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, P1489
  • [8] Cummins R., 2016, P 54 ANN M ASS COMP, V1
  • [9] Dauphin YN, 2015, ADV NEUR IN, V28
  • [10] Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171