Manipulation of sources of bias in AI device development

被引:0
作者
Burgon, Alexis [1 ]
Zhang, Yuhang [1 ]
Sahiner, Berkman [1 ]
Petrick, Nicholas [1 ]
Cha, Kenny H. [1 ]
Samala, Ravi K. [1 ]
机构
[1] US FDA, Ctr Devices & Radiol Hlth, Off Sci & Engn Labs, 10903 New Hampshire Ave, Silver Spring, MD 20993 USA
来源
COMPUTER-AIDED DIAGNOSIS, MEDICAL IMAGING 2024 | 2024年 / 12927卷
基金
美国国家卫生研究院;
关键词
artificial intelligence; bias; bias mitigation; fairness; image classification; machine learning;
D O I
10.1117/12.3008267
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The increased use of artificial intelligence (AI)-enabled medical devices in clinical practice has driven the need for better understanding of bias in AI applications. While there are many bias mitigation methods, determining which method is appropriate in each situation is not trivial due to a lack of objective bias mitigation assessment approaches. This study presents an approach to manipulate sources of bias in order to facilitate the evaluation of bias mitigation methods during AI development. This approach amplifies sources of bias in the data by varying disease prevalence between subgroups to promote associations between patient subgroups and specific disease classes. The approach can be adjusted to affect the degree to which bias is amplified. In this study, bias amplification was used to vary the bias of COVID-19 classification models using chest X-ray data between patient subgroups defined by sex (female or male) or race (black or white). Analysis of subgroup sensitivity shows that the proposed bias amplification approach results in the sensitivity of one subgroup increasing while the sensitivity of the other subgroup is decreasing, despite a consistent difference in subgroup area under the receiver operating characteristic curve. For example, the amplification of sex-related bias increased the sensitivity for COVID-19 classification of the female subgroup from 0.65 +/- 0.04 to 0.90 +/- 0.02 while decreasing the sensitivity of the male subgroup from 0.66 +/- 0.03 to 0.18 +/- 0.02. This shows that the bias amplification approaches promote bias by causing the model to enhance existing correlations between a patient's subgroup and their COVID-19 status, which results in biased performance in a clinical setting. The approach presented in this study was found to systematically amplify sources of bias for our dataset and AI model, and can thus be used to evaluate the effectiveness of potential bias mitigation methods.
引用
收藏
页数:6
相关论文
共 15 条
  • [1] Decision region analysis for generalizability of artificial intelligence models: estimating model generalizability in the case of cross-reactivity and population shift
    Burgon, Alexis
    Sahiner, Berkman
    Petrick, Nicholas
    Pennello, Gene
    Cha, Kenny H.
    Samala, Ravi K.
    [J]. JOURNAL OF MEDICAL IMAGING, 2024, 11 (01)
  • [2] Caton S, 2020, ACM Computing Surveys
  • [3] Deep Learning in Medical Image Analysis
    Chan, Heang-Ping
    Samala, Ravi K.
    Hadjiiski, Lubomir M.
    Zhou, Chuan
    [J]. DEEP LEARNING IN MEDICAL IMAGE ANALYSIS: CHALLENGES AND APPLICATIONS, 2020, 1213 : 3 - 21
  • [4] A Comprehensive Empirical Study of Bias Mitigation Methods for Machine Learning Classifiers
    Chen, Zhenpeng
    Zhang, Jie M.
    Sarro, Federica
    Harman, Mark
    [J]. ACM TRANSACTIONS ON SOFTWARE ENGINEERING AND METHODOLOGY, 2023, 32 (04)
  • [5] Proceedings of the NHLBI Workshop on Artificial Intelligence in Cardiovascular Imaging Translation to Patient Care
    Dey, Damini
    Arnaout, Rima
    Antani, Sameer
    Badano, Aldo
    Jacques, Louis
    Li, Huiqing
    Leiner, Tim
    Margerrison, Edward
    Samala, Ravi
    Sengupta, Partho P.
    Shah, Sanjiv J.
    Slomka, Piotr
    Williams, Michelle C.
    Bandettini, Patricia
    Sachdev, Vandana
    [J]. JACC-CARDIOVASCULAR IMAGING, 2023, 16 (09) : 1209 - 1223
  • [6] Toward fairness in artificial intelligence for medical image analysis: identification and mitigation of potential biases in the roadmap from data collection to model deployment
    Drukker, Karen
    Chen, Weijie
    Gichoya, Judy
    Gruszauskas, Nicholas
    Kalpathy-Cramer, Jayashree
    Koyejo, Sanmi
    Myers, Kyle
    Sa, Rui C.
    Sahiner, Berkman
    Whitney, Heather
    Zhang, Zi
    Giger, Maryellen
    [J]. JOURNAL OF MEDICAL IMAGING, 2023, 10 (06)
  • [7] Al recognition of patient race in medical imaging: a modelling study
    Gichoya, Judy Wawira
    Banerjee, Imon
    Bhimireddy, Ananth Reddy
    Burns, John L.
    Celi, Leo Anthony
    Chen, Li-Ching
    Correa, Ramon
    Dullerud, Natalie
    Ghassemi, Marzyeh
    Huang, Shih-Cheng
    Kuo, Po-Chih
    Lungren, Matthew P.
    Palmer, Lyle J.
    Price, Brandon J.
    Purkayastha, Saptarshi
    Pyrros, Ayis T.
    Oakden-Rayner, Lauren
    Okechukwu, Chima
    Seyyed-Kalantari, Laleh
    Trivedi, Hari
    Wang, Ryan
    Zaiman, Zachary
    Zhang, Haoran
    [J]. LANCET DIGITAL HEALTH, 2022, 4 (06): : E406 - E414
  • [8] Gossmann A., 2024, Deep Learning for Medical Image Analysis, P473
  • [9] AAPM task group report 273: Recommendations on best practices for AI and machine learning for computer-aided diagnosis in medical imaging
    Hadjiiski, Lubomir
    Cha, Kenny
    Chan, Heang-Ping
    Drukker, Karen
    Morra, Lia
    Nappi, Janne J.
    Sahiner, Berkman
    Yoshida, Hiroyuki
    Chen, Quan
    Deserno, Thomas M.
    Greenspan, Hayit
    Huisman, Henkjan
    Huo, Zhimin
    Mazurchuk, Richard
    Petrick, Nicholas
    Regge, Daniele
    Samala, Ravi
    Summers, Ronald M.
    Suzuki, Kenji
    Tourassi, Georgia
    Vergara, Daniel
    Armato, Samuel G., III
    [J]. MEDICAL PHYSICS, 2023, 50 (02) : E1 - E24
  • [10] Hort M., 2023, ACM J. Responsib. Comput.