Accounting for data variability in multi-institutional distributed deep learning for medical imaging

被引:43
作者
Balachandar, Niranjan [1 ]
Chang, Ken [2 ]
Kalpathy-Cramer, Jayashree [2 ,3 ]
Rubin, Daniel L. [1 ]
机构
[1] Stanford Univ, Lab Quantitat Imaging & Artificial Intelligence, Dept Radiol & Biomed Data Sci, Stanford, CA 94305 USA
[2] Massachusetts Gen Hosp, Athinoula A Martinos Ctr Biomed Imaging, Dept Radiol, Charlestown, MA USA
[3] Massachusetts Gen Hosp, MGH & BWH Ctr Clin Data Sci, Boston, MA 02114 USA
基金
美国国家卫生研究院;
关键词
distributed learning; federated learning; deep learning; medical imaging; transfer learning;
D O I
10.1093/jamia/ocaa017
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Objectives: Sharing patient data across institutions to train generalizable deep learning models is challenging due to regulatory and technical hurdles. Distributed learning, where model weights are shared instead of patient data, presents an attractive alternative. Cyclical weight transfer (CWT) has recently been demonstrated as an effective distributed learning method for medical imaging with homogeneous data across institutions. In this study, we optimize CWT to overcome performance losses from variability in training sample sizes and label distributions across institutions. Materials and Methods: Optimizations included proportional local training iterations, cyclical learning rate, locally weighted minibatch sampling, and cyclically weighted loss. We evaluated our optimizations on simulated distributed diabetic retinopathy detection and chest radiograph classification. Results: Proportional local training iteration mitigated performance losses from sample size variability, achieving 98.6% of the accuracy attained by centrally hosting in the diabetic retinopathy dataset split with highest sample size variance across institutions. Locally weighted minibatch sampling and cyclically weighted loss both mitigated performance losses from label distribution variability, achieving 98.6% and 99.1%, respectively, of the accuracy attained by centrally hosting in the diabetic retinopathy dataset split with highest label distribution variability across institutions. Discussion: Our optimizations to CWT improve its capability of handling data variability across institutions. Compared to CWT without optimizations, CWT with optimizations achieved performance significantly closer to performance from centrally hosting. Conclusion: Our work is the first to identify and address challenges of sample size and label distribution variability in simulated distributed deep learning for medical imaging. Future work is needed to address other sources of real-world data variability.
引用
收藏
页码:700 / 708
页数:9
相关论文
共 26 条
[1]   Deep learning for segmentation of brain tumors: Impact of cross-institutional training and testing [J].
AlBadawy, Ehab A. ;
Saha, Ashirbani ;
Mazurowski, Maciej A. .
MEDICAL PHYSICS, 2018, 45 (03) :1150-1158
[2]  
[Anonymous], C COMP VIS PATT REC
[3]  
[Anonymous], P 31 AAAI C
[4]  
[Anonymous], 2016, ARXIV E PRINTS
[5]  
[Anonymous], DIAB RET DET
[6]  
[Anonymous], 2015, ARXIV150701239
[7]  
[Anonymous], 2017, P IEEE C COMP VIS PA
[8]  
[Anonymous], IEEE T PATTERN ANAL
[9]  
[Anonymous], 2015, University of Warwick
[10]  
Bakas S., 2018, ARXIV181102629