Emotional speech analysis using harmonic plus noise model and Gaussian mixture model

被引:0
|
作者
Jang Bahadur Singh
Parveen Kumar Lehana
机构
[1] University of Jammu,DSP Lab, Department of Electronics
来源
International Journal of Speech Technology | 2019年 / 22卷
关键词
Harmonic plus noise model; Gaussian mixture model; Expectation maximization; Akaike information criterion;
D O I
暂无
中图分类号
学科分类号
摘要
Extracting the valuable information from the emotional speech is one of the major challenges in the areas of emotion recognition and human-machine interfaces. Most of the research in emotion recognition is based on the analysis of fundamental frequency, energy contour, duration of silence, formant, Mel-band energies, linear prediction cepstral coefficients, and Mel frequency cepstral coefficients. It was observed that emotion classification using sinusoidal features perform better as compared to the linear prediction and cepstral features. Harmonic models are considered as a variant of the sinusoidal model. In order to improve emotional speech classification rate and conversion of neutral speech to emotional speech, analysis using different harmonic features of emotional speech is a critical step. In this paper, investigations have been carried out using Berlin emotional speech database to analyze gender-based emotional speech using harmonic plus noise model (HNM) features and Gaussian mixture model (GMM). Analysis has been performed with the HNM features like pitch, harmonic amplitude, maximum voiced frequency and noise components. From the results, it can be observed that different emotional speech of male and female speakers can be represented with K components of GMM distribution. The optimal number of GMM components have been decided on the basis of Akaike information criterion (AIC) score.
引用
收藏
页码:483 / 496
页数:13
相关论文
共 50 条
  • [1] Emotional speech analysis using harmonic plus noise model and Gaussian mixture model
    Singh, Jang Bahadur
    Lehana, Parveen Kumar
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2019, 22 (03) : 483 - 496
  • [2] Enhancement of esophagus speech using harmonic plus noise model
    Lehana, PK
    Gupta, RK
    Kumari, S
    TENCON 2004 - 2004 IEEE REGION 10 CONFERENCE, VOLS A-D, PROCEEDINGS: ANALOG AND DIGITAL TECHNIQUES IN ELECTRICAL ENGINEERING, 2004, : A669 - A672
  • [3] A Voice Conversion System Based on the Harmonic plus Noise Excitation and Gaussian Mixture Model
    Wu Lifang
    Zhang Linghua
    PROCEEDINGS OF THE 2012 SECOND INTERNATIONAL CONFERENCE ON INSTRUMENTATION & MEASUREMENT, COMPUTER, COMMUNICATION AND CONTROL (IMCCC 2012), 2012, : 1575 - 1578
  • [4] Speech synthesis method with a harmonic plus noise model
    Ishikawa, Y
    Maruyama, I
    Hase, T
    2002 INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS, DIGEST OF TECHNICAL PAPERS, 2002, : 238 - 239
  • [5] Improving the modeling of the noise part in the Harmonic plus Noise model of speech
    Pantazis, Yannis
    Stylianou, Yannis
    2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4609 - +
  • [6] On the implementation of the Harmonic plus Noise Model for concatenative speech synthesis
    Stylianou, Y
    2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS, VOLS I-VI, 2000, : 957 - 960
  • [7] Applying the harmonic plus noise model in concatenative speech synthesis
    Stylianou, Y
    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, 2001, 9 (01): : 21 - 29
  • [8] Classification of stressed speech using Gaussian mixture model
    Patro, H
    Raja, GS
    Dandapat, S
    INDICON 2005 Proceedings, 2005, : 342 - 346
  • [9] ANALYSIS/SYNTHESIS OF SPEECH BASED ON AN ADAPTIVE QUASI-HARMONIC PLUS NOISE MODEL
    Pantazis, Yannis
    Tzedakis, Georgios
    Rosec, Olivier
    Stylianou, Yannis
    2010 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2010, : 4246 - 4249
  • [10] A Long-Term Harmonic plus Noise Model for Speech Signals
    Ben Ali, Faten
    Girin, Laurent
    Larbi, Sonia Djaziri
    12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 60 - +