Exploring AI Techniques for Generalizable Teaching Practice Identification

被引:0
作者
Garcia, Federico Pardo [1 ]
Canovas, Oscar [1 ]
Clemente, Felix J. Garcia [1 ]
机构
[1] Univ Murcia, Dept Ingn & Tecnol Comp, Murcia 30100, Spain
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Audio analysis; deep learning; machine learning; multi-modal learning analytics; speaker diarization; teaching practices; CLASSIFICATION;
D O I
10.1109/ACCESS.2024.3456915
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Using automated models to analyze classroom discourse is a valuable tool for educators to improve their teaching methods. In this paper, we focus on exploring alternatives to ensure the generalizability of models for identifying teaching practices across diverse teaching contexts. Our proposal utilizes artificial intelligence to analyze audio recordings of classroom activities. By leveraging deep learning for speaker diarization and traditional machine learning algorithms for classifying teaching practices, we extract features from the audio diarization using a processing pipeline to provide detailed insights into teaching dynamics. These features enable the classification of three distinct teaching practices: lectures, group discussions, and the use of audience response systems. Our findings demonstrate that these features effectively capture the nuances of teacher-student interactions, allowing for a refined analysis of teaching styles. To enhance the robustness and generalizability of our model, we explore various pipelines for audio processing, evaluating the model's performance across diverse contexts involving different teachers and students. By comparing these practices and their associated features, we illustrate how AI-driven tools can support teachers in reflecting on and improving their teaching strategies.
引用
收藏
页码:134702 / 134713
页数:12
相关论文
共 35 条
  • [1] Multimodal Speaker Diarization Using a Pre-Trained Audio-Visual Synchronization Model
    Ahmad, Rehan
    Zubair, Syed
    Alquhayz, Hani
    Ditta, Allah
    [J]. SENSORS, 2019, 19 (23)
  • [2] Archer J, 2016, JosseyBass
  • [3] A Multimodal-Sensor-Enabled Room for Unobtrusive Group Meeting Analysis
    Bhattacharya, Indrani
    Foley, Michael
    Zhang, Ni
    Zhang, Tongtao
    Ku, Christine
    Mine, Cameron
    Ji, Heng
    Riedl, Christoph
    Welles, Brooke Foucault
    Radke, Richard J.
    [J]. ICMI'18: PROCEEDINGS OF THE 20TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2018, : 347 - 355
  • [4] Bibal A., 2016, EUR S ART NEUR NETW, P77
  • [5] Blikstein P., 2016, J LEARN ANAL, V3, P220, DOI [DOI 10.18608/JLA.2016.32.11, 10.18608/jla.2016.32.11]
  • [6] Bredin H, 2020, INT CONF ACOUST SPEE, P7124, DOI [10.1109/icassp40776.2020.9052974, 10.1109/ICASSP40776.2020.9052974]
  • [7] Buitinck Lars, 2013, P EUR C MACH LEARN P, P108
  • [8] Canovas O., 2023, P IEEE INT C TEACH A, P1
  • [9] Measuring the effect of ARS on academic performance: A global meta-analysis
    Castillo-Manzano, Jose I.
    Castro-Nuno, Mercedes
    Lopez-Valpuesta, Lourdes
    Teresa Sanz-Diaz, Maria
    Yniguez, Rocio
    [J]. COMPUTERS & EDUCATION, 2016, 96 : 109 - 121
  • [10] Chejara Pankaj, 2023, LAK2023: LAK23: 13th International Learning Analytics and Knowledge Conference, P111, DOI 10.1145/3576050.3576144