Cross-modal Transfer Learning Based on an Improved CycleGAN Model for Accurate Kidney Segmentation in Ultrasound Images

被引:1
|
作者
Guo, Shuaizi [1 ]
Chen, Haijie [1 ]
Sheng, Xiangyu [1 ]
Xiong, Yinzheng [2 ]
Wu, Menglin [1 ,3 ]
Fischer, Katherine [4 ,5 ]
Tasian, Gregory E. [4 ,5 ,6 ]
Fan, Yong [7 ]
Yin, Shi [1 ]
机构
[1] Nanjing Tech Univ, Sch Comp Sci & Technol, Nanjing 211816, Peoples R China
[2] Northeastern Univ, Khoury Coll Comp Sci, Boston, MA USA
[3] Carbon Med Device Ltd, Shenzhen, Peoples R China
[4] Childrens Hosp Philadelphia, Dept Surg, Div Pediat Urol, Philadelphia, PA USA
[5] Childrens Hosp Philadelphia, Ctr Pediat Clin Effectiveness, Philadelphia, PA USA
[6] Univ Penn, Dept Biostat Epidemiol & Informat, Philadelphia, PA USA
[7] Univ Penn, Perelman Sch Med, Dept Radiol, Philadelphia, PA USA
来源
ULTRASOUND IN MEDICINE AND BIOLOGY | 2024年 / 50卷 / 11期
基金
美国国家卫生研究院;
关键词
Kidney US images; Segmentation network; CycleGAN model; CT dataset; Transfer learning; LEVEL SET; MEDICAL ULTRASOUND; DISTANCE; SIMULATION;
D O I
10.1016/j.ultrasmedbio.2024.06.009
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Objective: Deep-learning algorithms have been widely applied in the field of automatic kidney ultrasound (US) image segmentation. However, obtaining a large number of accurate kidney labels clinically is very difficult and time-consuming. To solve this problem, we have proposed an efficient cross-modal transfer learning method to improve the performance of the segmentation network on a limited labeled kidney US dataset. Methods: We aim to implement an improved image-to-image translation network called Seg-CycleGAN to generate accurate annotated kidney US data from labeled abdomen computed tomography images. The Seg-CycleGAN framework primarily consists of two structures: (i) a standard CycleGAN network to visually simulate kidney US from a publicly available labeled abdomen computed tomography dataset; (ii) and a segmentation network to ensure accurate kidney anatomical structures in US images. Based on the large number of simulated kidney US images and small number of real annotated kidney US images, we then aimed to employ a fine-tuning strategy to obtain better segmentation results. Results: To validate the effectiveness of the proposed method, we tested this method on both normal and abnormal kidney US images. The experimental results showed that the proposed method achieved a segmentation accuracy of 0.8548 in dice similarity coefficient on all testing datasets and 0.7622 on the abnormal testing dataset. Conclusions: Compared with existing data augmentation and transfer learning methods, the proposed method improved the accuracy and generalization of the kidney US image segmentation network on a limited number of training datasets. It therefore has the potential to significantly reduce annotation costs in clinical settings.
引用
收藏
页码:1638 / 1645
页数:8
相关论文
共 50 条
  • [1] CT2US: Cross-modal transfer learning for kidney segmentation in ultrasound images with synthesized data
    Song, Yuxin
    Zheng, Jing
    Lei, Long
    Ni, Zhipeng
    Zhao, Baoliang
    Hu, Ying
    ULTRASONICS, 2022, 122
  • [2] Road segmentation of cross-modal remote sensing images using deep segmentation network and transfer learning
    He, Hao
    Yang, Dongfang
    Wang, Shicheng
    Wang, Shuyang
    Liu, Xing
    INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2019, 46 (03): : 384 - 390
  • [3] A Classification Method for the Cellular Images Based on Active Learning and Cross-Modal Transfer Learning
    Vununu, Caleb
    Lee, Suk-Hwan
    Kwon, Ki-Ryong
    SENSORS, 2021, 21 (04) : 1 - 24
  • [4] MCKTNet: Multiscale Cross-Modal Knowledge Transfer Network for Semantic Segmentation of Remote Sensing Images
    Cui, Jian
    Liu, Jiahang
    Ni, Yue
    Sun, Yuan
    Guo, Mao
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2025, 63
  • [5] An incremental cross-modal transfer learning method for gesture interaction
    Zhong, Junpei
    Li, Jie
    Lotfi, Ahmad
    Liang, Peidong
    Yang, Chenguang
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2022, 155
  • [6] Transfer Learning Based Fully Automated Kidney Segmentation on MR Images
    Gaikar, Rohini
    Zabihollahy, Fatemeh
    Farrag, Nadia
    Elfaal, Mohamed W.
    Schieda, Nicola
    Ukwatta, Eranga
    MEDICAL IMAGING 2022: BIOMEDICAL APPLICATIONS IN MOLECULAR, STRUCTURAL, AND FUNCTIONAL IMAGING, 2022, 12036
  • [7] Cross-Modal Learning for Event-Based Semantic Segmentation via Attention Soft Alignment
    Xie, Chuyun
    Gao, Wei
    Guo, Ren
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (03): : 2359 - 2366
  • [8] Transfer learning for cross-modal demand prediction of bike-share and public transit
    Hua, Mingzhuang
    Pereira, Francisco Camara
    Jiang, Yu
    Chen, Xuewu
    Chen, Junyi
    JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS, 2024,
  • [9] Cross-modal hashing network based on self-attention similarity transfer
    Liang H.
    Wang H.
    Wang D.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (02): : 615 - 622
  • [10] Classification of Alzheimer Disease on Imaging Modalities with Deep CNNs using Cross-Modal Transfer Learning
    Aderghal, Karim
    Khvostikov, Alexander
    Krylov, Andrei
    Benois-Pineau, Jenny
    Afdel, Karim
    Catheline, Gwenaelle
    2018 31ST IEEE INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS 2018), 2018, : 345 - 350