Articulatory-Acoustic Analyses of Mandarin Words in Emotional Context Speech for Smart Campus

被引:10
作者
Ren, Guofeng [1 ,2 ]
Zhang, Xueying [1 ]
Duan, Shufei [1 ]
机构
[1] Taiyuan Univ Technol, Coll Informat & Comp, Jinzhong 030600, Peoples R China
[2] Xizhou Teachers Univ, Dept Elect, Xinzhou 034000, Peoples R China
关键词
Articulatory-acoustic analysis; electromagnetic articulography; emotional speech processing; Mandarin word; smart campus; ELECTROMAGNETIC ARTICULOGRAPHY; KINEMATICS; MOVEMENTS; VOWELS; LEVEL;
D O I
10.1109/ACCESS.2018.2865831
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recent years, along with the promotion of smart campus, social networks developed rapidly which demands high accuracy of man-man & man-machine interaction technologies. Thus, physiological information in speech interactive processing has become an important complement or even replaced acoustic-based features. With the aim of assessing the influence of emotions on articulatory-acoustic features in speech production, the current study explored the articulatory mechanism that underlying emotional speech production of Mandarin words. We first used the AG501 EMA device to collect articulatory and acoustic data synchronously as subjects were speaking specific words in Mandarin with different emotions, e.g., anger, sadness, happiness, and neutral; articulatory and acoustic features then were extracted from the collected data and analyzed in a one-way ANOVA to discover the significance of emotions on articulatory and acoustic features. The results illustrated that the motion of articulators (tongue and lip) were influenced by emotions significantly; in detail, the motion range of tongue and lip with anger were larger than other emotions, meanwhile, tongue speed and lip speed with anger and happiness were more sensitive than with sadness and neutral in emotional words. Results had been discussed to discover the relationship between acoustic and articulatory features of emotional speech, and then the conclusion can be acquired that articulatory motion feature (tongue and lip) may be the major feature of emotional speech recognition, so that which can be applied to the man-machine interaction of smart campus research in the future.
引用
收藏
页码:48418 / 48427
页数:10
相关论文
共 31 条
[1]  
[Anonymous], P MSR BING IRC 2013
[2]  
Chilela J. G., 2016, 2014184162 U COIMBR
[3]  
Cowie R., 2000, PROC ISCA WORKSHOP S, P19
[4]   Comparison of emotion perception among different cultures [J].
Dang, Jianwu ;
Li, Aijun ;
Erickson, Donna ;
Suemitsu, Atsuo ;
Akagi, Masato ;
Sakuraba, Kyoko ;
Minematsu, Nobuaki ;
Hirose, Keikichi .
ACOUSTICAL SCIENCE AND TECHNOLOGY, 2010, 31 (06) :394-402
[5]   Articulation, Acoustics and Perception of Mandarin Chinese Emotional Speech [J].
Erickson, Donna ;
Zhu, Chunyue ;
Kawahara, Shigeto ;
Suemitsu, Atsuo .
OPEN LINGUISTICS, 2016, 2 (01) :620-635
[6]   The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing [J].
Eyben, Florian ;
Scherer, Klaus R. ;
Schuller, Bjoern W. ;
Sundberg, Johan ;
Andre, Elisabeth ;
Busso, Carlos ;
Devillers, Laurence Y. ;
Epps, Julien ;
Laukka, Petri ;
Narayanan, Shrikanth S. ;
Truong, Khiet P. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2016, 7 (02) :190-202
[7]  
Heracleous P, 2011, INT CONF ACOUST SPEE, P2392
[8]   How fluent is the fluent speech of people who stutter? A new approach to measuring kinematics with ultrasound [J].
Heyde, Cornelia J. ;
Scobbie, James M. ;
Lickley, Robin ;
Drake, Eleanor K. E. .
CLINICAL LINGUISTICS & PHONETICS, 2016, 30 (3-5) :292-312
[9]   A kinematic study of critical and non-critical articulators in emotional speech production [J].
Kim, Jangwon ;
Toutios, Asterios ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 2015, 137 (03) :1411-1429
[10]   Automatic intelligibility classification of sentence-level pathological speech [J].
Kim, Jangwon ;
Kumar, Naveen ;
Tsiartas, Andreas ;
Li, Ming ;
Narayanan, Shrikanth S. .
COMPUTER SPEECH AND LANGUAGE, 2015, 29 (01) :132-144