Data and model bias in artificial intelligence for healthcare applications in New Zealand

被引:6
作者
Yogarajan, Vithya [1 ]
Dobbie, Gillian [1 ]
Leitch, Sharon [2 ]
Keegan, Te Taka [3 ]
Bensemann, Joshua [1 ]
Witbrock, Michael [1 ]
Asrani, Varsha [4 ]
Reith, David [5 ]
机构
[1] Univ Auckland, Sch Comp Sci, Waipapa Taumata Rau, Auckland, New Zealand
[2] Univ Otago, Otago Med Sch, Gen Practice & Rural Hlth, Dunedin, New Zealand
[3] Univ Waikato, Sch Comp & Math Sci, Hamilton, New Zealand
[4] Univ Auckland, Fac Med & Hlth Sci, Surg & Translat Res STaR Ctr, Sch Med,Dept Surg, Auckland, New Zealand
[5] Univ Otago, Otago Med Sch, Dunedin, New Zealand
来源
FRONTIERS IN COMPUTER SCIENCE | 2022年 / 4卷
关键词
Artificial Intelligence; bias; healthcare; New Zealand; Maori; equity; CONCEPT DRIFT;
D O I
10.3389/fcomp.2022.1070493
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
IntroductionDevelopments in Artificial Intelligence (AI) are adopted widely in healthcare. However, the introduction and use of AI may come with biases and disparities, resulting in concerns about healthcare access and outcomes for underrepresented indigenous populations. In New Zealand, Maori experience significant inequities in health compared to the non-Indigenous population. This research explores equity concepts and fairness measures concerning AI for healthcare in New Zealand. MethodsThis research considers data and model bias in NZ-based electronic health records (EHRs). Two very distinct NZ datasets are used in this research, one obtained from one hospital and another from multiple GP practices, where clinicians obtain both datasets. To ensure research equality and fair inclusion of Maori, we combine expertise in Artificial Intelligence (AI), New Zealand clinical context, and te ao Maori. The mitigation of inequity needs to be addressed in data collection, model development, and model deployment. In this paper, we analyze data and algorithmic bias concerning data collection and model development, training and testing using health data collected by experts. We use fairness measures such as disparate impact scores, equal opportunities and equalized odds to analyze tabular data. Furthermore, token frequencies, statistical significance testing and fairness measures for word embeddings, such as WEAT and WEFE frameworks, are used to analyze bias in free-form medical text. The AI model predictions are also explained using SHAP and LIME. ResultsThis research analyzed fairness metrics for NZ EHRs while considering data and algorithmic bias. We show evidence of bias due to the changes made in algorithmic design. Furthermore, we observe unintentional bias due to the underlying pre-trained models used to represent text data. This research addresses some vital issues while opening up the need and opportunity for future research. DiscussionsThis research takes early steps toward developing a model of socially responsible and fair AI for New Zealand's population. We provided an overview of reproducible concepts that can be adopted toward any NZ population data. Furthermore, we discuss the gaps and future research avenues that will enable more focused development of fairness measures suitable for the New Zealand population's needs and social structure. One of the primary focuses of this research was ensuring fair inclusions. As such, we combine expertise in AI, clinical knowledge, and the representation of indigenous populations. This inclusion of experts will be vital moving forward, proving a stepping stone toward the integration of AI for better outcomes in healthcare.
引用
收藏
页数:21
相关论文
共 68 条
[1]  
[Anonymous], 2017, International Journal on Advances in Software, V10, P1
[2]   The intensivist's assessment of gastrointestinal function: A pilot study [J].
Asrani, Varsha M. ;
McArthur, Colin ;
Bissett, Ian ;
Windsor, John A. .
AUSTRALIAN CRITICAL CARE, 2022, 35 (06) :636-643
[3]  
Badilla P, 2020, PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P430
[4]   Prediction across healthcare settings: a case study in predicting emergency department disposition [J].
Barak-Corren, Yuval ;
Chaudhari, Pradip ;
Perniciaro, Jessica ;
Waltzman, Mark ;
Fine, Andrew M. ;
Reis, Ben Y. .
NPJ DIGITAL MEDICINE, 2021, 4 (01)
[5]  
Besse P., 2022, AM STAT, V76, P188, DOI [10.1080/00031305.2021.1952897, DOI 10.1080/00031305.2021.1952897]
[6]  
Bisong E., 2019, Logistic regression. Building machine learning and deep learning models on google cloud platform: a comprehensive guide for beginners, P243
[7]  
Burlot F, 2017, TRAIT AUTOM LANG, V58, P98
[8]   Semantics derived automatically from language corpora contain human-like biases [J].
Caliskan, Aylin ;
Bryson, Joanna J. ;
Narayanan, Arvind .
SCIENCE, 2017, 356 (6334) :183-186
[9]   Machine Learning Interpretability: A Survey on Methods and Metrics [J].
Carvalho, Diogo, V ;
Pereira, Eduardo M. ;
Cardoso, Jaime S. .
ELECTRONICS, 2019, 8 (08)
[10]   Sources of bias in artificial intelligence that perpetuate healthcare disparities-A global review [J].
Celi, Leo Anthony ;
Cellini, Jacqueline ;
Charpignon, Marie-Laure ;
Dee, Edward Christopher ;
Dernoncourt, Franck ;
Eber, Rene ;
Mitchell, William Greig ;
Moukheiber, Lama ;
Schirmer, Julian ;
Situ, Julia ;
Paguio, Joseph ;
Park, Joel ;
Wawira, Judy Gichoya ;
Yao, Seth .
PLOS DIGITAL HEALTH, 2022, 1 (03)