Segment anything model for medical images?

被引:165
作者
Huang, Yuhao [1 ,2 ,3 ]
Yang, Xin [1 ,2 ,3 ]
Liu, Lian [1 ,2 ,3 ]
Zhou, Han [1 ,2 ,3 ]
Chang, Ao [1 ,2 ,3 ]
Zhou, Xinrui [1 ,2 ,3 ]
Chen, Rusi [1 ,2 ,3 ]
Yu, Junxuan [1 ,2 ,3 ]
Chen, Jiongquan [1 ,2 ,3 ]
Chen, Chaoyu [1 ,2 ,3 ]
Liu, Sijing [1 ,2 ,3 ]
Chi, Haozhe [2 ,4 ]
Hu, Xindi [2 ,5 ]
Yue, Kejuan [2 ,6 ]
Li, Lei [2 ,7 ]
Grau, Vicente [2 ,7 ]
Fan, Deng-Ping [2 ,8 ]
Dong, Fajin [2 ,3 ,9 ,10 ]
Ni, Dong [1 ,2 ,3 ]
机构
[1] Shenzhen Univ, Med Sch, Sch Biomed Engn, Natl Reg Key Technol Engn Lab Med Ultrasound, Shenzhen 518060, Peoples R China
[2] Shenzhen Univ, Med Ultrasound Image Comp MUS Lab, Shenzhen, Peoples R China
[3] Shenzhen Univ, Marshall Lab Biomed Engn, Shenzhen, Peoples R China
[4] Zhejiang Univ, Zhejiang, Peoples R China
[5] Shenzhen RayShape Med Technol Co Ltd, Shenzhen, Peoples R China
[6] Hunan First Normal Univ, Changsha, Peoples R China
[7] Univ Oxford, Dept Engn Sci, Oxford, England
[8] Swiss Fed Inst Technol, Comp Vis Lab CVL, Zurich, Switzerland
[9] Jinan Univ, Clin Med Coll 2, Ultrasound Dept, Guangzhou, Peoples R China
[10] Southern Univ Sci & Technol, Affiliated Hosp 1, Shenzhen Peoples Hosp, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
Segment anything model; Medical image segmentation; Medical object perception; ALGORITHMS; VALIDATION; FRAMEWORK;
D O I
10.1016/j.media.2023.103061
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Segment Anything Model (SAM) is the first foundation model for general image segmentation. It has achieved impressive results on various natural image segmentation tasks. However, medical image segmenta-tion (MIS) is more challenging because of the complex modalities, fine anatomical structures, uncertain and complex object boundaries, and wide-range object scales. To fully validate SAM's performance on medical data, we collected and sorted 53 open-source datasets and built a large medical segmentation dataset with 18 modalities, 84 objects, 125 object-modality paired targets, 1050K 2D images, and 6033K masks. We comprehensively analyzed different models and strategies on the so-called COSMOS 1050K dataset. Our findings mainly include the following: (1) SAM showed remarkable performance in some specific objects but was unstable, imperfect, or even totally failed in other situations. (2) SAM with the large ViT-H showed better overall performance than that with the small ViT-B. (3) SAM performed better with manual hints, especially box, than the Everything mode. (4) SAM could help human annotation with high labeling quality and less time. (5) SAM was sensitive to the randomness in the center point and tight box prompts, and may suffer from a serious performance drop. (6) SAM performed better than interactive methods with one or a few points, but will be outpaced as the number of points increases. (7) SAM's performance correlated to different factors, including boundary complexity, intensity differences, etc. (8) Finetuning the SAM on specific medical tasks could improve its average DICE performance by 4.39% and 6.68% for ViT-B and ViT-H, respectively. Codes and models are available at: https://github.com/yuhoo0302/Segment-Anything-Model-for-Medical-Images. We hope that this comprehensive report can help researchers explore the potential of SAM applications in MIS, and guide how to appropriately use and develop SAM.
引用
收藏
页数:21
相关论文
共 101 条
[91]   A deep learning based medical image segmentation technique in Internet-of-Medical-Things domain [J].
Wang, Eric Ke ;
Chen, Chien-Ming ;
Hassan, Mohammad Mehedi ;
Almogren, Ahmad .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 108 :135-144
[92]  
Wang Ziyi, 2022, arXiv
[93]   TotalSegmentator: Robust Segmentation of 104 Anatomic Structures in CT Images [J].
Wasserthal, Jakob ;
Breit, Hanns-Christian ;
Meyer, Manfred T. ;
Pradella, Maurice ;
Hinck, Daniel ;
Sauter, Alexander W. ;
Heye, Tobias ;
Boll, Daniel T. ;
Cyriac, Joshy ;
Yang, Shan ;
Bach, Michael ;
Segeroth, Martin .
RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2023, 5 (05)
[94]  
Williams DR., 2021, The confidence interval that wasn't: bootstrapped confidence intervals in l1regularized partial correlation networks
[95]  
Wu JD, 2023, Arxiv, DOI [arXiv:2304.12620, 10.48550/arXiv.2304.12620, DOI 10.48550/ARXIV.2304.12620]
[96]   Colonoscopic image synthesis with generative adversarial network for enhanced detection of sessile serrated lesions using convolutional neural network [J].
Yoon, Dan ;
Kong, Hyoun-Joong ;
Kim, Byeong Soo ;
Cho, Woo Sang ;
Lee, Jung Chan ;
Cho, Minwoo ;
Lim, Min Hyuk ;
Yang, Sun Young ;
Lim, Seon Hee ;
Lee, Jooyoung ;
Song, Ji Hyun ;
Chung, Goh Eun ;
Choi, Ji Min ;
Kang, Hae Yeon ;
Bae, Jung Ho ;
Kim, Sungwan .
SCIENTIFIC REPORTS, 2022, 12 (01)
[97]   Robust Retinal Vessel Segmentation via Locally Adaptive Derivative Frames in Orientation Scores [J].
Zhang, Jiong ;
Dashtbozorg, Behdad ;
Bekkers, Erik ;
Pluim, Josien P. W. ;
Duits, Remco ;
Romeny, Bart M. ter Haar .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2016, 35 (12) :2631-2644
[98]   A Coarse-to-Fine Framework for the 2021 Kidney and Kidney Tumor Segmentation Challenge [J].
Zhao, Zhongchen ;
Chen, Huai ;
Wang, Lisheng .
KIDNEY AND KIDNEY TUMOR SEGMENTATION, KITS 2021, 2022, 13168 :53-58
[99]   IoU Loss for 2D/3D Object Detection [J].
Zhou, Dingfu ;
Fang, Jin ;
Song, Xibin ;
Guan, Chenye ;
Yin, Junbo ;
Dai, Yuchao ;
Yang, Ruigang .
2019 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2019), 2019, :85-94
[100]  
Zhou T, 2023, Arxiv, DOI arXiv:2304.07583