Aspect-Based Semantic Textual Similarity for Educational Test Items

被引:0
作者
Do, Heejin [1 ]
Lee, Gary Geunbae [1 ,2 ]
机构
[1] POSTECH, Grad Sch AI, Pohang, South Korea
[2] POSTECH, Dept CSE, Pohang, South Korea
来源
ARTIFICIAL INTELLIGENCE IN EDUCATION, PT II, AIED 2024 | 2024年 / 14830卷
关键词
Educational Item Similarity; Semantic Textual Similarity; Dataset; Aspect-based Similarity; Natural Language Processing;
D O I
10.1007/978-3-031-64299-9_30
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the educational domain, identifying the similarity among test items provides various advantages for exam quality management and personalized student learning. Existing studies mostly relied on student performance data, such as the number of correct or incorrect answers, to measure item similarity. However, nuanced semantic information within the test items has been overlooked, possibly due to the lack of similarity-labeled data. Human-annotated educational data demands high-cost expertise, and items comprising multiple aspects, such as questions and choices, require detailed criteria. In this paper, we introduce a task of aspect-based semantic textual similarity for educational test items (aSTS-EI), where we assess the similarity by specific aspects within test items and present an LLM-guided benchmark dataset. We report the baseline performance by extending the STS methods, setting the groundwork for future aSTS-EI tasks. In addition, to assist data-scarce settings, we propose a progressive augmentation (ProAug) method, which generates step-by-step item aspects via recursive prompting. Experimental results imply the efficacy of existing STS methods for a shorter aspect while underlining the necessity for specialized approaches in relatively longer aspects. Nonetheless, markedly improved results with ProAug highlight the assistance of our augmentation strategy to overcome data scarcity.
引用
收藏
页码:344 / 352
页数:9
相关论文
共 24 条
  • [1] Brown TB, 2020, ADV NEUR IN, V33
  • [2] Cechak J., 2021, Intern. Educ. Data Mining Soc.
  • [3] Cer Daniel, 2017, P 11 INT WORKSHOP SE, P1, DOI 10.18653/v1/S17-2001
  • [4] Chiang CH, 2023, Arxiv, DOI [arXiv:2305.01937, DOI 10.48550/ARXIV.2305.01937]
  • [5] Chung Yu-An, 2018, NAACL HLT
  • [6] From Distillation to Hard Negative Sampling: Making Sparse Neural IR Models More Effective
    Formal, Thibault
    Lassance, Carlos
    Piwowarski, Benjamin
    Clinchant, Stephane
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2353 - 2359
  • [7] Fu JL, 2023, Arxiv, DOI [arXiv:2302.04166, DOI 10.48550/ARXIV.2302.04166]
  • [8] Gao TY, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P6894
  • [9] Geng C, 2021, J. Phys. Conf. Ser.
  • [10] Harbouche K, 2023, Ingenierie des Systemes d'Information