Bias Invariant Approaches for Improving Word Embedding Fairness

被引:1
作者
Liao, Siyu [1 ]
Zhang, Rongting [1 ]
Poblete, Barbara [1 ,2 ]
Murdock, Vanessa [1 ]
机构
[1] Amazon Com, Seattle, WA 98109 USA
[2] Univ Chile, Santiago, Chile
来源
PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023 | 2023年
关键词
Word Embedding; Fairness; Privacy;
D O I
10.1145/3583780.3614792
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many public pre-trained word embeddings have been shown to encode different types of biases. Embeddings are often obtained from training on large pre-existing corpora, and therefore resulting biases can be a reflection of unfair representations in the original data. Bias, in this scenario, is a challenging problem since current mitigation techniques require knowing and understanding existing biases in the embedding, which is not always possible. In this work, we propose to improve word embedding fairness by borrowing methods from the field of data privacy. The idea behind this approach is to treat bias as if it were a special type of training data leakage. This has the unique advantage of not requiring prior knowledge of potential biases in word embeddings. We investigated two types of privacy algorithms, and measured their effect on bias using four different metrics. To investigate techniques from differential privacy, we applied Gaussian perturbation to public pre-trained word embeddings. To investigate noiseless privacy, we applied vector quantization during training. Experiments show that both approaches improve fairness for commonly used embeddings, and additionally, noiseless privacy techniques reduce the size of the resulting embedding representation.
引用
收藏
页码:1400 / 1410
页数:11
相关论文
共 48 条
[1]   Deep Learning with Differential Privacy [J].
Abadi, Martin ;
Chu, Andy ;
Goodfellow, Ian ;
McMahan, H. Brendan ;
Mironov, Ilya ;
Talwar, Kunal ;
Zhang, Li .
CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, :308-318
[2]  
Badilla P, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P430
[3]  
Balle B, 2018, PR MACH LEARN RES, V80
[4]  
Barocas S., 2020, P 58 ANN M ASS COMP, P5454, DOI DOI 10.18653/V1/2020.ACL-MAIN.485
[5]  
Bengio Y, 2001, ADV NEUR IN, V13, P932
[6]  
BHASKAR R, 2011, INT C THEOR APPL CRY, V7073, P215
[7]  
Bolukbasi T, 2016, ADV NEUR IN, V29
[8]   Semantics derived automatically from language corpora contain human-like biases [J].
Caliskan, Aylin ;
Bryson, Joanna J. ;
Narayanan, Arvind .
SCIENCE, 2017, 356 (6334) :183-186
[9]  
Chiu Billy, 2016, P 1 WORKSHOP EVALUAT, P1, DOI DOI 10.18653/V1/W16-2501
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171