Stationary wavelet Filtering Cepstral coefficients (SWFCC) for robust speaker identification

被引:0
|
作者
Missaoui, Ibrahim [1 ,2 ]
Lachiri, Zied [1 ]
机构
[1] Univ Tunis El Manar, Natl Engn Sch Tunis ENIT, Signal Images & Informat Technol Lab, LR-11-ES17,BP 37, Tunis 1002, Tunisia
[2] Univ Gabes, Higher Inst Comp Sci & Multimedia Gabes, Gabes, Tunisia
关键词
Stationary wavelet filtering cepstral; coefficients; SWFCC; SWT; Stationary wavelet packet transform; Implicit wiener filtering; Feature extraction; GMM-UBM; Robust speaker recognition; SPEECH WAVE; PACKET;
D O I
10.1016/j.apacoust.2024.110435
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Extracting robust effective speech features is one of the challenging topics in the speaker recognition field, especially in noisy conditions. It can substantially improve the robustness recognition accuracy of persons from their voice signals against such conditions. This paper proposes a new feature extraction approach called Stationary Wavelet Filtering Cepstral Coefficients (SWFCC) for noisy speaker recognition. The proposed approach incorporates a Stationary Wavelet Filterbank (SWF) and an Implicit Wiener Filtering (IWF) technique. The SWF is based on the stationary wavelet packet transform, which is a shift-invariant transform. The performance of the proposed SWFCC approach is evaluated on the TIMIT dataset in the presence of different types of environmental noise, which are taken from the Aurora dataset. Our experimental results using the Gaussian Mixture ModelUniversal Background Model (GMM-UBM) as a classifier show that SWFCC outperforms various feature extraction techniques like MFCC, PNCC, and GFCC in terms of recognition accuracy.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] A robust speaker identification system based on wavelet transform
    Hsieh, CT
    Wang, YC
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2001, E84D (07): : 839 - 846
  • [22] Speaker identification using cepstral analysis
    Nazar, MN
    ISCON 2002: IEEE STUDENTS CONFERENCE ON EMERGING TECHNOLOGIES, PROCEEDINGS, 2002, : 139 - 143
  • [23] Speaker Identification Using Linear Predictive Cepstral Coefficients And General Regression Neural Network
    Li, Penghua
    Hu, Fangchao
    Li, Yinguo
    Xu, Yang
    2014 33RD CHINESE CONTROL CONFERENCE (CCC), 2014, : 4952 - 4956
  • [24] Wavelet Packet Based Mel Frequency Cepstral Features for Text Independent Speaker Identification
    Srivastava, Smriti
    Bhardwaj, Saurabh
    Bhandari, Abhishek
    Gupta, Krit
    Bahl, Hitesh
    Gupta, J. R. P.
    INTELLIGENT INFORMATICS, 2013, 182 : 237 - 247
  • [25] Power Normalized Gammachirp Cepstral (PNGC) coefficients-based approach for robust speaker recognition
    Zouhir, Youssef
    Zarka, Mohamed
    Supervision, Kais Ouni
    APPLIED ACOUSTICS, 2023, 205
  • [26] Wavelet Cepstral Coefficients for Electrical Appliances Identification using Hidden Markov Models
    Hacine-Gharbi, Abdenour
    Ravier, Philippe
    PROCEEDINGS OF THE 7TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS (ICPRAM 2018), 2018, : 541 - 549
  • [27] Perceptual MVDR-based cepstral coefficients for speaker recognition
    Liang, Chunyan
    Zhang, Xiang
    Yang, Lin
    Zhang, Jianping
    Yan, Yonghong
    Shengxue Xuebao/Acta Acustica, 2012, 37 (06): : 673 - 678
  • [28] Mel Frequency Cepstral Coefficients (MFCC) Based Speaker Identification in Noisy Environment Using Wiener Filter
    Chauhan, Paresh M.
    Desai, Nikita P.
    2014 INTERNATIONAL CONFERENCE ON GREEN COMPUTING COMMUNICATION AND ELECTRICAL ENGINEERING (ICGCCEE), 2014,
  • [29] Robust speech features based on wavelet transform with application to speaker identification
    Hsieh, CT
    Lai, E
    Wang, YC
    IEE PROCEEDINGS-VISION IMAGE AND SIGNAL PROCESSING, 2002, 149 (02): : 108 - 114
  • [30] A Study on Speaker Identification Approach by Feature Matching Algorithm using Pitch and Mel Frequency Cepstral Coefficients
    Prasetio, Barlian Henryranu
    Sakurai, Keiko
    Tamura, Hiroki
    Tanno, Koichi
    ICAROB 2019: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON ARTIFICIAL LIFE AND ROBOTICS, 2019, : 475 - 478