VERB: Visualizing and Interpreting Bias Mitigation Techniques Geometrically forWord Representations

被引:3
作者
Rathore, Archit [1 ]
Dev, Sunipa [2 ]
Phillips, Jeff M. [1 ]
Srikumar, Vivek [1 ]
Zheng, Yan [3 ]
Yeh, Chin-Chia Michael [3 ]
Wang, Junpeng [3 ]
Zhang, Wei [3 ]
Wang, Bei [1 ]
机构
[1] Univ Utah, Salt Lake City, UT 84112 USA
[2] Univ Calif Los Angeles, Los Angeles, CA USA
[3] VISA Res, Palo Alto, CA USA
基金
美国国家科学基金会;
关键词
Responsible XAI; debiasing; visual data exploration; distributed representations; bias mitigation; model interpretation; ethics;
D O I
10.1145/3604433
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Word vector embeddings have been shown to contain and amplify biases in the data they are extracted from. Consequently, many techniques have been proposed to identify, mitigate, and attenuate these biases in word representations. In this article, we utilize interactive visualization to increase the interpretability and accessibility of a collection of state-of-the-art debiasing techniques. To aid this, we present the Visualization of Embedding Representations for deBiasing (VERB) system, an open-source web-based visualization tool that helps users gain a technical understanding and visual intuition of the inner workings of debiasing techniques, with a focus on their geometric properties. In particular, VERB offers easy-to-followexamples that explore the effects of these debiasing techniques on the geometry of high-dimensional word vectors. To help understand how various debiasing techniques change the underlying geometry, VERB decomposes each technique into interpretable sequences of primitive transformations and highlights their effect on the word vectors using dimensionality reduction and interactive visual exploration. VERB is designed to target natural language processing (NLP) practitioners who are designing decision-making systems on top of word embeddings and researchers working with the fairness and ethics of machine learning systems in NLP. It can also serve as a visual medium for education, which helps an NLP novice understand and mitigate biases in word embeddings.
引用
收藏
页数:34
相关论文
共 81 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   ModelTracker: Redesigning Performance Analysis Tools for Machine Learning [J].
Amershi, Saleema ;
Chickering, Max ;
Drucker, Steven M. ;
Lee, Bongshin ;
Simard, Patrice ;
Suh, Jina .
CHI 2015: PROCEEDINGS OF THE 33RD ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2015, :337-346
[3]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[4]  
Blodgett SL., 2020, NLP, P5454, DOI [10.18653/v1/2020.acl-main.485, DOI 10.18653/V1/2020.ACL-MAIN.485]
[5]  
Bolukbasi T, 2016, ADV NEUR IN, V29
[6]  
Brown TB, 2020, ADV NEUR IN, V33
[7]  
Cabrera AA, 2019, IEEE CONF VIS ANAL, P46, DOI [10.1109/VAST47406.2019.8986948, 10.1109/vast47406.2019.8986948]
[8]   Semantics derived automatically from language corpora contain human-like biases [J].
Caliskan, Aylin ;
Bryson, Joanna J. ;
Narayanan, Arvind .
SCIENCE, 2017, 356 (6334) :183-186
[9]   OoDAnalyzer: Interactive Analysis of Out-of-Distribution Samples [J].
Chen, Changjian ;
Yuan, Jun ;
Lu, Yafeng ;
Liu, Yang ;
Su, Hang ;
Yuan, Songtao ;
Liu, Shixia .
IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, 2021, 27 (07) :3335-3349
[10]   Visual Analytics for Explainable Deep Learning [J].
Choo, Jaegul ;
Liu, Shixia .
IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2018, 38 (04) :84-92