Gender-sensitive word embeddings for healthcare

被引:4
作者
Agmon, Shunit [1 ]
Gillis, Plia [2 ]
Horvitz, Eric [3 ]
Radinsky, Kira [1 ]
机构
[1] Technion Israel Inst Technol, Comp Sci Fac, Haifa, Israel
[2] Tel Aviv Univ, Fac Med, Tel Aviv, Israel
[3] Microsoft Res, Redmond, WA USA
关键词
word embeddings; statistical models; bias; algorithms; gender; PREVALENCE; WOMEN; UMLS;
D O I
10.1093/jamia/ocab279
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objective To analyze gender bias in clinical trials, to design an algorithm that mitigates the effects of biases of gender representation on natural-language (NLP) systems trained on text drawn from clinical trials, and to evaluate its performance. Materials and Methods We analyze gender bias in clinical trials described by 16 772 PubMed abstracts (2008-2018). We present a method to augment word embeddings, the core building block of NLP-centric representations, by weighting abstracts by the number of women participants in the trial. We evaluate the resulting gender-sensitive embeddings performance on several clinical prediction tasks: comorbidity classification, hospital length of stay prediction, and intensive care unit (ICU) readmission prediction. Results For female patients, the gender-sensitive model area under the receiver-operator characteristic (AUROC) is 0.86 versus the baseline of 0.81 for comorbidity classification, mean absolute error 4.59 versus the baseline of 4.66 for length of stay prediction, and AUROC 0.69 versus 0.67 for ICU readmission. All results are statistically significant. Discussion Women have been underrepresented in clinical trials. Thus, using the broad clinical trials literature as training data for statistical language models could result in biased models, with deficits in knowledge about women. The method presented enables gender-sensitive use of publications as training data for word embeddings. In experiments, the gender-sensitive embeddings show better performance than baseline embeddings for the clinical tasks studied. The results highlight opportunities for recognizing and addressing gender and other representational biases in the clinical trials literature. Conclusion Addressing representational biases in data for training NLP embeddings can lead to better results on downstream tasks for underrepresented populations.
引用
收藏
页码:415 / 423
页数:9
相关论文
共 58 条
  • [1] [Anonymous], Women's Health Problems
  • [2] [Anonymous], 20 TESTICULAR DISORD
  • [3] [Anonymous], 2018, NAT HLTH INT SURV
  • [4] [Anonymous], INSIDE EPIDEMIC MISD
  • [5] [Anonymous], WHAT ARE AUTOIMMUNE
  • [6] [Anonymous], 2018, ARXIV180401486
  • [7] CodeMagic: Semi-Automatic Assignment of ICD-10-AM Codes to Patient Records
    Arifoglu, Damla
    Deniz, Onur
    Alecakir, Kemal
    Yondem, Meltem
    [J]. INFORMATION SCIENCES AND SYSTEMS 2014, 2014, : 259 - 268
  • [8] Aronson AR, 2001, J AM MED INFORM ASSN, P17
  • [9] Basta C, 2019, P 1 WORKSH GEND BIAS
  • [10] Data-Driven Decisions for Reducing Readmissions for Heart Failure: General Methodology and Case Study
    Bayati, Mohsen
    Braverman, Mark
    Gillam, Michael
    Mack, Karen M.
    Ruiz, George
    Smith, Mark S.
    Horvitz, Eric
    [J]. PLOS ONE, 2014, 9 (10):