Speaker recognition based on short utterance compensation method of generative adversarial networks

被引:0
作者
Zhangfang Hu
Yaqin Fu
Yuan Luo
Xuan Xu
Zhiguang Xia
Hongwei Zhang
机构
[1] Chongqing University of Posts and Telecommunications,College of Optoelectronic Engineering
来源
International Journal of Speech Technology | 2020年 / 23卷
关键词
Gaussian mixture model–universal background model; Speaker recognition; Generative adversarial network;
D O I
暂无
中图分类号
学科分类号
摘要
On the basis of gaussian mixture model–universal background model (GMM–UBM) in the speaker recognition system, the paper proposes a short utterance sample compensation method based on the generative adversarial network (GAN) to solve the problem of the inadequate corpus data caused by short utterance, which has led to a serious reduction of recognition rate. The presented method compensates the short utterance samples into the speech samples with sufficient speaker identity information by completing the antagonistic training of generator network and discriminator network. In order to avoid the model crash and gradient instability in the process of GAN training, this paper adopts the condition information in the conditional GAN to guide the compensation process of the generator network, and proposes the generator compensation performance measurement training task and the feature tag training task of the discriminator to stabilize training process. Finally, the proposed short utterance compensation method is evaluated on the speaker recognition system based on GMM–UBM. The experimental results indicate that the presented method can effectively reduce the equal error rate of the speaker recognition system in short utterance environment.
引用
收藏
页码:443 / 450
页数:7
相关论文
共 50 条
  • [31] Mixed-type data generation method based on generative adversarial networks
    Ning Wei
    Longzhi Wang
    Guanhua Chen
    Yirong Wu
    Shunfa Sun
    Peng Chen
    EURASIP Journal on Wireless Communications and Networking, 2022
  • [32] Multi-resolution time frequency feature and complementary combination for short utterance speaker recognition
    Zhi-Yi Li
    Wei-Qiang Zhang
    Jia Liu
    Multimedia Tools and Applications, 2015, 74 : 937 - 953
  • [33] Mixed-type data generation method based on generative adversarial networks
    Wei, Ning
    Wang, Longzhi
    Chen, Guanhua
    Wu, Yirong
    Sun, Shunfa
    Chen, Peng
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2022, 2022 (01)
  • [34] Length- and Noise-aware Training Techniques for Short-utterance Speaker Recognition
    Chen, Wenda
    Huang, Jonathan
    Bocklet, Tobias
    INTERSPEECH 2020, 2020, : 3835 - 3839
  • [35] Multi-resolution time frequency feature and complementary combination for short utterance speaker recognition
    Li, Zhi-Yi
    Zhang, Wei-Qiang
    Liu, Jia
    MULTIMEDIA TOOLS AND APPLICATIONS, 2015, 74 (03) : 937 - 953
  • [36] Enhanced Iris Recognition Method by Generative Adversarial Network-Based Image Reconstruction
    Lee, Min Beom
    Kang, Jin Kyu
    Yoon, Hyo Sik
    Park, Kang Ryoung
    IEEE ACCESS, 2021, 9 : 10120 - 10135
  • [37] Data Augmentation Using Deep Generative Models for Embedding Based Speaker Recognition
    Wang, Shuai
    Yang, Yexin
    Wu, Zhanghao
    Qian, Yanmin
    Yu, Kai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 2598 - 2609
  • [38] Adversarial Examples Detection for XSS Attacks Based on Generative Adversarial Networks
    Zhang, Xueqin
    Zhou, Yue
    Pei, Songwen
    Zhuge, Jingjing
    Chen, Jiahao
    IEEE ACCESS, 2020, 8 (08): : 10989 - 10996
  • [39] Generative Adversarial Networks-Based Semi-Supervised Automatic Modulation Recognition for Cognitive Radio Networks
    Li, Mingxuan
    Li, Ou
    Liu, Guangyi
    Zhang, Ce
    SENSORS, 2018, 18 (11)
  • [40] An anomaly detection method based on double encoder-decoder generative adversarial networks
    Liu, Hui
    Tang, Tinglong
    Luo, Jake
    Zhao, Meng
    Zheng, Baole
    Wu, Yirong
    INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2021, 48 (05): : 643 - 648