AI-driven mock interview assessment: leveraging generative language models for automated evaluation

被引:1
作者
Uppalapati, Padma Jyothi [1 ,2 ]
Dabbiru, Madhavi [3 ]
Kasukurthi, Venkata Rao [4 ]
机构
[1] Andhra Univ, Andhra Univ Transdisciplinary Res Hub, Dept CS&SE, Coll Engn, Visakhapatnam 530003, AP, India
[2] Vishnu Inst Technol, Dept CSE, Bhimavaram 534202, AP, India
[3] Dr Lankapalli Bullayya Coll, Dept CSE, Visakhapatnam 530013, AP, India
[4] Andhra Univ, Dept CS&SE, Coll Engn, Visakhapatnam 530003, AP, India
关键词
Videos of mock interview; Generated language models; Automatic speech recognition; Speaker diarization; Technical skills; WhisperX;
D O I
10.1007/s13042-025-02529-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the education sector, adaptive support is critical for every student to face open-ended activities that need behavioral change, performance, and a pro-active learning mindset. This can be accomplished by using brilliant learning environments powered by artificial intelligence. Timely feedback is critical for helping students enhance their overall personality in learning, confidence, communication, and problem-solving. It is a highly demanding task for every teacher to conduct a mock interview and then provide feedback for each criterion. It is time consuming and may be delayed. However automated review of mock interviews can give timely student feedback while reducing the manual evaluation burden on teachers in areas with a high teacher-to-student ratio. Current ways of analyzing student interview responses include transformer-based natural language processing models, which have various degrees of effectiveness. One major problem in training these models the need of more data, as most of the datasets are based on HR queries, which have sufficient datasets. But, none of the interview recordings included both HR and TR-related questions. We recorded mock interviews with undergraduate students from multiple backgrounds to address the data scarcity issue. We extracted audio from the recordings, followed by transcripts that included speaker identification. We split the speakers' data into questions and responses. The biggest challenge is evaluating these answers, given the need for appropriate datasets for technical questions. This article investigates the text-generating AI model, GPT-3.5, to establish whether prompt-based text-generation approaches are viable for generating scores for specific student responses. Finally model results are compared with human values. Our results reveal that the pre-trained model yields excellent outcomes in interview grading.
引用
收藏
页数:23
相关论文
empty
未找到相关数据