A multifaceted approach to detect gender biases in Natural Language Generation

被引:0
作者
Consuegra-Ayala, Juan Pablo [1 ]
Martinez-Murillo, Ivan [3 ]
Lloret, Elena [2 ,3 ]
Moreda, Paloma [2 ,3 ]
Palomar, Manuel [2 ,3 ]
机构
[1] Univ Havana, Sch Math & Comp Sci, Havana 10200, Cuba
[2] Univ Alicante, Univ Inst Comp Res IUII, Alicante 03690, Spain
[3] Univ Alicante, Dept Language & Comp Syst, Alicante 03690, Spain
关键词
Natural Language Generation; Gender bias; Common sense resources; RISK;
D O I
10.1016/j.knosys.2024.112367
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent advances in generative models have skyrocketed the popularity of conversational chatbots and have revolutionized the way we interact with artificial intelligence. At the same time, research has shown that machine learning models can unconsciously reflect and amplify human biases. This is particularly dangerous for generative models given the huge popularity of such technologies. Specifically, a fundamental source of bias of such technologies is the resources on which the models are trained. To address this issue, this paper proposes a methodology to analyze intrinsic gender bias in Natural Language Generation (NLG). Some works already propose metrics and approaches to measure bias in the Natural Language processing field. However, there is a lack of standard methodology to measure gender bias in NLG. Therefore, adapting the Bias Score approach, our proposal involves three sequential stages applied to individual texts to detect intrinsic gender bias on NLG effectively. Those steps are as follows: (i) word scoring; (ii) word filtering; and (iii) generative-word analysis. This methodology is applied to recent datasets and pre-trained models widely used for the generation of text with common sense. In particular, this paper analyzes the potential gender bias in CommonGen and C2Gen 2 Gen datasets and the SimpleNLG and T5 models. The results show the ability of the proposed methodology to detect gender bias in word distributions, presenting a strong correlation with the words typically associated with a specific gender. Results indicate that both tested datasets are intrinsically gender-biased, and therefore, tested models fine-tuned with those datasets also are.
引用
收藏
页数:12
相关论文
共 53 条
  • [21] Indurkhya N, 2010, Handbook of natural language processing, V2
  • [22] Kaneko M, 2022, AAAI CONF ARTIF INTE, P11954
  • [23] Consumer credit-risk models via machine-learning algorithms
    Khandani, Amir E.
    Kim, Adlar J.
    Lo, Andrew W.
    [J]. JOURNAL OF BANKING & FINANCE, 2010, 34 (11) : 2767 - 2787
  • [24] Applicants' Fairness Perceptions of Algorithm-Driven Hiring Procedures
    Lavanchy, Maude
    Reichert, Patrick
    Narayanan, Jayanth
    Savani, Krishna
    [J]. JOURNAL OF BUSINESS ETHICS, 2023, 188 (01) : 125 - 150
  • [25] Improving the accuracy of top-N recommendation using a preference model
    Lee, Jongwuk
    Lee, Dongwon
    Lee, Yeon-Chang
    Hwang, Won-Seok
    Kim, Sang-Wook
    [J]. INFORMATION SCIENCES, 2016, 348 : 290 - 304
  • [26] Exploiting the potentialities of features for speech emotion recognition
    Li, Dongdong
    Zhou, Yijun
    Wang, Zhe
    Gao, Daqi
    [J]. INFORMATION SCIENCES, 2021, 548 : 328 - 343
  • [27] Neural Natural Language Processing for unstructured data in electronic health records: A review
    Li, Irene
    Pan, Jessica
    Goldwasser, Jeremy
    Verma, Neha
    Wong, Wai Pan
    Nuzumlali, Muhammed Yavuz
    Rosand, Benjamin
    Li, Yixin
    Zhang, Matthew
    Chang, David
    Taylor, R. Andrew
    Krumholz, Harlan M.
    Radev, Dragomir
    [J]. COMPUTER SCIENCE REVIEW, 2022, 46
  • [28] Lin BY, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, P1823
  • [29] Lin C.Y., 2004, ANN M ASS COMPUTATI, P74, DOI DOI 10.1253/JCJ.34.1213
  • [30] Mahoney J.F., 2007, US Patent, Patent No. [7,287,008, 7287008]