MFUnetr: A transformer-based multi-task learning network for multi-organ segmentation from partially labeled datasets

被引:9
作者
Hao, Qin [1 ,2 ]
Tian, Shengwei [1 ,2 ]
Yu, Long [3 ,4 ]
Wang, Junwen [1 ,2 ]
机构
[1] Xinjiang Univ, Coll Software, Urumqi 830000, Peoples R China
[2] Xinjiang Univ, Key Lab Software Engn Technol, Urumqi 830000, Peoples R China
[3] Xinjiang Univ, Coll Network Ctr, Urumqi 830000, Peoples R China
[4] Xinjiang Univ, Coll Informat Sci & Engn, Signal & Signal Proc Lab, Urumqi 830000, Peoples R China
基金
中国国家自然科学基金;
关键词
Partially labeled dataset; 3D CT image segmentation; Multi -task learning; Multi -organ segmentation;
D O I
10.1016/j.bspc.2023.105081
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
As multi-organ segmentation of CT images is crucial for clinical applications, most state-of-the-art models rely on a fully annotated dataset with strong supervision to achieve high accuracy for particular organs. However, these models have weak generalization when applied to various CT images due to the small scale and single source of the training data. To utilize existing partially labeled datasets to obtain segmentation containing more organs and with higher accuracy and robustness, we create a multi-task learning network called MFUnetr. By directly feeding a union of datasets, MFUnetr trains an encoder-decoder network on two tasks in parallel. The main task is to produce full organ segmentation using a specific training strategy. The auxiliary task is to segment labeled organs of each dataset using label priors. Additionally, we offer a new weighted combined loss function to optimize the model. Compared to the base model UNETR trained on the fully annotated dataset BTCV, our network model, utilizing a combination of three partially labeled datasets, achieved mean Dice on overlapping organs: spleen + 0.35 %, esophagus + 15.28 %, and aorta + 8.31 %. Importantly, without fine-tuning, the mean Dice calculated on 13 organs of BTCV remained + 1.91 % when all 15 organs were segmented. The experimental results show that our proposed method can effectively use existing large partially annotated datasets to alleviate the problem of data scarcity in multi-organ segmentation.
引用
收藏
页数:15
相关论文
共 52 条
[1]   The Lovasz-Softmax loss: A tractable surrogate for the optimization of the intersection-over-union measure in neural networks [J].
Berman, Maxim ;
Triki, Amal Rannen ;
Blaschko, Matthew B. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4413-4421
[2]   Pancreas segmentation by two-view feature learning and multi-scale supervision [J].
Chen, Haipeng ;
Liu, Yunjie ;
Shi, Zenan ;
Lyu, Yingda .
BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2022, 74
[3]  
Chen J., 2021, arXiv, DOI DOI 10.48550/ARXIV.2102.04306
[4]   Physical Structure Induced Hydrophobicity Analyzed from Electrospinning and Coating Polyvinyl Butyral Films [J].
Chen, Shuo ;
Liu, Guo-Sai ;
He, Hong-Wei ;
Zhou, Cheng-Feng ;
Yan, Xu ;
Zhang, Jun-Cheng .
ADVANCES IN CONDENSED MATTER PHYSICS, 2019, 2019
[5]  
Cicek Ozgun, 2016, Medical Image Computing and Computer-Assisted Intervention - MICCAI 2016. 19th International Conference. Proceedings: LNCS 9901, P424, DOI 10.1007/978-3-319-46723-8_49
[6]   Learning Multi-Class Segmentations From Single-Class Datasets [J].
Dmitriev, Konstantin ;
Kaufman, Arie E. .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9493-9503
[7]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[8]   Multi-Organ Segmentation Over Partially Labeled Datasets With Multi-Scale Feature Abstraction [J].
Fang, Xi ;
Yan, Pingkun .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (11) :3619-3629
[9]   Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images [J].
Graham, Simon ;
Quoc Dang Vu ;
Raza, Shan E. Ahmed ;
Azam, Ayesha ;
Tsang, Yee Wah ;
Kwak, Jin Tae ;
Rajpoot, Nasir .
MEDICAL IMAGE ANALYSIS, 2019, 58
[10]   Dynamic Task Prioritization for Multitask Learning [J].
Guo, Michelle ;
Haque, Albert ;
Huang, De-An ;
Yeung, Serena ;
Li Fei-Fei .
COMPUTER VISION - ECCV 2018, PT XVI, 2018, 11220 :282-299