AI-based English teaching cross-cultural fusion mechanism

被引:5
作者
Wang, Fang [1 ]
机构
[1] Huanghe Sci & Technol Univ, Zhengzhou 450063, Peoples R China
关键词
English teaching; Convolutional neural networks; Cross-cultural; Spectrogram; Speech emotion recognition; EMOTION RECOGNITION; ARTIFICIAL-INTELLIGENCE; EDUCATION;
D O I
10.1007/s12065-022-00733-y
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the differences between eastern and western cultures, many students often face difficulties in language learning caused by cultural differences in learning English. Adaptability in cross-cultural communication and fluency in language learning are the highest realm of learning English and the highest pursuit for every student's learning. To push the dialogue contents for students, this paper proposes a recognition model of multi-input feature extraction based on convolutional neural networks (CNN) for English emotion recognition. The model adopts an end-to-end training method for learning, which can simplify the workflow and improve the overall recognition accuracy. At first, a speech is transformed into a form more suitable for model learning: spectrogram and Mel spectrogram. Then, it is input into the model for automatic feature extraction. In the model design, the information of spectrogram in time-domain and frequency-domain is considered at the same time. Finally, the high-level features of the image are fused, which greatly improves the effect of speech emotion recognition. Experimental results prove that the proposed speech emotion recognition fusion model has high recognition accuracy and quality of experience.
引用
收藏
页码:1461 / 1467
页数:7
相关论文
共 26 条
  • [1] Technology-Assisted Self-Regulated English Language Learning: Associations With English Language Self-Efficacy, English Enjoyment, and Learning Outcomes
    An, Zhujun
    Wang, Chuang
    Li, Siying
    Gan, Zhengdong
    Li, Hong
    [J]. FRONTIERS IN PSYCHOLOGY, 2021, 11
  • [2] Improved speech emotion recognition with Mel frequency magnitude coefficient
    Ancilin, J.
    Milton, A.
    [J]. APPLIED ACOUSTICS, 2021, 179
  • [3] Deep features-based speech emotion recognition for smart affective services
    Badshah, Abdul Malik
    Rahim, Nasir
    Ullah, Noor
    Ahmad, Jamil
    Muhammad, Khan
    Lee, Mi Young
    Kwon, Soonil
    Baik, Sung Wook
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (05) : 5571 - 5589
  • [4] English teaching practice based on artificial intelligence technology
    Bin, Yi
    Mandal, Durbadal
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 37 (03) : 3381 - 3391
  • [5] Cong Meng, 2020, Journal of Physics: Conference Series, V1648, DOI 10.1088/1742-6596/1648/2/022036
  • [6] Impact of Approximate Multipliers on VGG Deep Learning Network
    Hammad, Issam
    El-Sankary, Kamal
    [J]. IEEE ACCESS, 2018, 6 : 60438 - 60444
  • [7] Hou M, 2022, IEEE-ACM T AUDIO SPE, V30
  • [8] Understanding English language learning in tertiary English-medium instruction contexts in China
    Hu, Jingjing
    Wu, Peng
    [J]. SYSTEM, 2020, 93
  • [9] Huang X, 2021, AIMS CULTIVATING STU, V26
  • [10] Speech emotion recognition using emotion perception spectral feature
    Jiang, Lin
    Tan, Ping
    Yang, Junfeng
    Liu, Xingbao
    Wang, Chao
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2021, 33 (11)