Adversarial Reweighting for Speaker Verification Fairness

被引:2
作者
Jin, Minho [1 ]
Ju, Chelsea J-T [2 ]
Chen, Zeya [2 ]
Liu, Yi-Chieh [2 ]
Droppo, Jasha [2 ]
Stolcke, Andreas [2 ]
机构
[1] Amazon Web Serv, Palo Alto, CA 94303 USA
[2] Amazon Alexa AI, Sunnyvale, CA USA
来源
INTERSPEECH 2022 | 2022年
关键词
speaker verification; speaker recognition; fairness; RECOGNITION;
D O I
10.21437/Interspeech.2022-10948
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We address performance fairness for speaker verification using the adversarial reweighting (ARW) method. ARW is reformulated for speaker verification with metric learning, and shown to improve results across different subgroups of gender and nationality, without requiring annotation of subgroups in the training data. An adversarial network learns a weight for each training sample in the batch so that the main learner is forced to focus on poorly performing instances. Using a min-max optimization algorithm, this method improves overall speaker verification fairness. We present three different ARW formulations: accumulated pairwise similarity, pseudo-labeling, and pairwise weighting, and measure their performance in terms of equal error rate (EER) on the VoxCeleb corpus. Results show that the pairwise weighting method can achieve 1.08% overall EER, 1.25% for male and 0.67% for female speakers, with relative EER reductions of 7.7%, 10.1% and 3.0%, respectively. For nationality subgroups, the proposed algorithm showed 1.04% EER for US speakers, 0.76% for UK speakers, and 1.22% for all others. The absolute EER gap between gender groups was reduced from 0.70% to 0.58%, while the standard deviation over nationality groups decreased from 0.21 to 0.19.
引用
收藏
页码:4800 / 4804
页数:5
相关论文
共 30 条
  • [11] Kamiran Faisal, 2010, PROC 19 MACHINE LEAR, P1
  • [12] Racial disparities in automated speech recognition
    Koenecke, Allison
    Nam, Andrew
    Lake, Emily
    Nudell, Joe
    Quartey, Minnie
    Mengesha, Zion
    Toups, Connor
    Rickford, John R.
    Jurafsky, Dan
    Goel, Sharad
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2020, 117 (14) : 7684 - 7689
  • [13] LAHOTI P, 2020, ADV NEUR IN, V33, pNI524
  • [14] MODEL-BASED APPROACH FOR MEASURING THE FAIRNESS IN ASR
    Liu, Zhe
    Veliche, Irina-Elena
    Peng, Fuchun
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6532 - 6536
  • [15] MacQueen J, 1967, P 5 BERK S MATH STAT, V1, P281, DOI DOI 10.1007/S11665-016-2173-6
  • [16] A Survey on Bias and Fairness in Machine Learning
    Mehrabi, Ninareh
    Morstatter, Fred
    Saxena, Nripsuta
    Lerman, Kristina
    Galstyan, Aram
    [J]. ACM COMPUTING SURVEYS, 2021, 54 (06)
  • [17] Menon A. K., 2018, C FAIRNESS ACCOUNTAB, P107
  • [18] VoxCeleb: a large-scale speaker identification dataset
    Nagrani, Arsha
    Chung, Joon Son
    Zisserman, Andrew
    [J]. 18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2616 - 2620
  • [19] Attentive Statistics Pooling for Deep Speaker Embedding
    Okabel, Koji
    Koshinaka, Takafumi
    Shinoda, Koichi
    [J]. 19TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2018), VOLS 1-6: SPEECH RESEARCH FOR EMERGING MARKETS IN MULTILINGUAL SOCIETIES, 2018, : 2252 - 2256
  • [20] Taking Advantage of Multitask Learning for Fair Classification
    Oneto, Luca
    Donini, Michele
    Elders, Amon
    Pontil, Massimiliano
    [J]. AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, : 227 - 237