Duration Dependent Covariance Regularization in PLDA Modeling for Speaker Verification

被引:0
|
作者
Cai, Weicheng [2 ,3 ]
Li, Ming [1 ,2 ]
Li, Lin [4 ]
Hong, Qingyang [4 ]
机构
[1] Sun Yat Sen Univ, SYSU CMU Joint Inst Engn, Guangzhou, Guangdong, Peoples R China
[2] SYSU CMU Shunde Int Joint Res Inst, Shunde, Guangdong, Peoples R China
[3] Sun Yat Sen Univ, Sch Informat Sci & Technol, Guangzhou, Guangdong, Peoples R China
[4] Xiamen Univ, Sch Informat Sci & Technol, Xiamen, Peoples R China
来源
16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5 | 2015年
关键词
PLDA; covariance regularization; i-vector; speaker verification; duration; ROBUST;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, we present a covariance regularized probabilistic linear discriminant analysis (CR-PLDA) model for text independent speaker verification. In the conventional simplified PLDA modeling, the covariance matrix used to capture the residual energies is globally shared for all i-vectors. However, we believe that the point estimated i-vectors from longer speech utterances may be more accurate and their corresponding co-variances in the PLDA modeling should be smaller. Similar to the inverse 0th order statistics weighted covariance in the i-vector model training, we propose a duration dependent normalized exponential term containing the duration normalizing factor mu and duration extent factor v to regularize the covariance in the PLDA modeling. Experimental results are reported on the NIST SRE 2010 common condition 5 female part task and the NIST 2014 i-vector machine learning challenge, respectively. For both tasks, the proposed covariance regularized PLDA system outperforms the baseline PLDA system by more than 13% relatively in terms of equal error rate (EER) and norm minDCF values.
引用
收藏
页码:1027 / 1031
页数:5
相关论文
共 50 条
  • [21] Unifying Cosine and PLDA Back-ends for Speaker Verification
    Peng, Zhiyuan
    He, Xuanji
    Ding, Ke
    Lee, Tan
    Wan, Guanglu
    INTERSPEECH 2022, 2022, : 336 - 340
  • [22] PLDA inspired Siamese networks for speaker verification
    Ramoji, Shreyas
    Krishnan, Prashant
    Ganapathy, Sriram
    COMPUTER SPEECH AND LANGUAGE, 2022, 76
  • [23] A TRANSFER LEARNING METHOD FOR PLDA-BASED SPEAKER VERIFICATION
    Hong, Qingyang
    Zhang, Jun
    Li, Lin
    Wan, Lihong
    Tong, Feng
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 5455 - 5459
  • [24] An Iterative Framework for Unsupervised Learning in the PLDA based Speaker Verification
    Liu, Wenbo
    Yu, Zhiding
    Li, Ming
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 78 - +
  • [25] MULTI-OBJECTIVE OPTIMIZATION TRAINING OF PLDA FOR SPEAKER VERIFICATION
    He, Liang
    Chen, Xianhong
    Xu, Can
    Liu, Jia
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6026 - 6030
  • [26] Performance Evaluation of Mixtures of PLDA and Conventional PLDA for a Small-Set Speaker Verification System
    Wan, Qianhui
    Bouchard, Martin
    2017 IEEE 30TH CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (CCECE), 2017,
  • [27] Robust speaker verification with state duration modeling
    Yoma, NB
    Pegoraro, TF
    SPEECH COMMUNICATION, 2002, 38 (1-2) : 77 - 88
  • [28] FULL-COVARIANCE UBM AND HEAVY-TAILED PLDA IN I-VECTOR SPEAKER VERIFICATION
    Matejka, Pavel
    Glembek, Ondrej
    Castaldo, Fabio
    Alam, M. J.
    Plchot, Oldrich
    Kenny, Patrick
    Burget, Lukas
    Cernocky, Jan 'Honza'
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 4828 - 4831
  • [29] Analysis of the Influence of Speech Corpora in the PLDA Verification in the Task of Speaker Recognition
    Machlica, Lukas
    Zajic, Zbynek
    TEXT, SPEECH AND DIALOGUE, TSD 2012, 2012, 7499 : 464 - 471
  • [30] Scoring of Large-Margin Embeddings for Speaker Verification: Cosine or PLDA?
    Wang, Qiongqiong
    Lee, Kong Aik
    Liu, Tianchi
    INTERSPEECH 2022, 2022, : 600 - 604