A Survey on Bias in Deep NLP

被引:64
作者
Garrido-Munoz, Ismael [1 ]
Montejo-Raez, Arturo [1 ]
Martinez-Santiago, Fernando [1 ]
Urena-Lopez, L. Alfonso [1 ]
机构
[1] Ctr Estudios Avanzados TIC CEATIC, Jaen 230071, Spain
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 07期
关键词
natural language processing; deep learning; biased models;
D O I
10.3390/app11073184
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Deep neural networks are hegemonic approaches to many machine learning areas, including natural language processing (NLP). Thanks to the availability of large corpora collections and the capability of deep architectures to shape internal language mechanisms in self-supervised learning processes (also known as "pre-training"), versatile and performing models are released continuously for every new network design. These networks, somehow, learn a probability distribution of words and relations across the training collection used, inheriting the potential flaws, inconsistencies and biases contained in such a collection. As pre-trained models have been found to be very useful approaches to transfer learning, dealing with bias has become a relevant issue in this new scenario. We introduce bias in a formal way and explore how it has been treated in several networks, in terms of detection and correction. In addition, available resources are identified and a strategy to deal with bias in deep NLP is proposed.
引用
收藏
页数:26
相关论文
共 87 条
  • [1] Abid Abubakar, 2021, ARXIV210105783
  • [2] [Anonymous], 1996, Text and Corpus Analysis: Computer-Assisted Studies of Language and Culture
  • [3] Antoniou G, 2003, IN HAND I S, P67
  • [4] Baader F, 2003, IN HAND I S, P3
  • [5] Quantifying Gender Bias in Different Corpora
    Babaeianjelodar, Marzieh
    Lorenz, Stephen
    Gordon, Josh
    Matthews, Jeanna
    Freitag, Evan
    [J]. WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 752 - 759
  • [6] Stereotypical Bias Removal for Hate Speech Detection Task using Knowledge-based Generalizations
    Badjatiya, Pinkesh
    Gupta, Manish
    Varma, Vasudeva
    [J]. WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 49 - 59
  • [7] Bartl M., 2020, P 2 WORKSH GEND BIAS, P1
  • [8] Basta C, 2019, GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019), P33
  • [9] Bender E. M., 2018, Transactions of the Association for Computational Linguistics, V6, P587, DOI [10.1162/tacla00041, DOI 10.1162/TACLA00041]
  • [10] Bethke A., 2020, StereoSet: Measuring stereotypical bias in pretrained language models