Multi-path decoder U-Net: A weakly trained real-time segmentation network for object detection and localization in ultrasound scans

被引:7
作者
Al-Battal, Abdullah F. [1 ,2 ]
Lerman, Imanuel R. [1 ,3 ]
Nguyen, Truong Q. [1 ]
机构
[1] Univ Calif San Diego, Elect & Comp Engn Dept, San Diego, CA 92093 USA
[2] King Fahd Univ Petr & Minerals, Elect Engn Dept, Dhahran, Saudi Arabia
[3] Univ Calif San Diego, UC San Diego Hlth, San Diego, CA 92093 USA
关键词
Deep convolutional neural networks; Ultrasound; Object detection; U-Net; LESIONS; IMAGES;
D O I
10.1016/j.compmedimag.2023.102205
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Detecting and localizing an anatomical structure of interest within the field of view of an ultrasound scan is an essential step in many diagnostic and therapeutic procedures. However, ultrasound scans suffer from high levels of variabilities across sonographers and patients, making it challenging for sonographers to accurately identify and locate these structures without extensive experience. Segmentation-based convolutional neural networks (CNNs) have been proposed as a solution to assist sonographers in this task. Despite their accuracy, these networks require pixel-wise annotations for training; an expensive and labor-intensive operation that requires the expertise of an experienced practitioner to identify the precise outline of the structures of interest. This complicates, delays, and increases the cost of network training and deployment. To address this problem, we propose a multi-path decoder U-Net architecture that is trained on bounding box segmentation maps; not requiring pixel-wise annotations. We show that the network can be trained on small training sets, which is the case in medical imaging datasets; reducing the cost and time needed for deployment and use in clinical settings. The multi-path decoder design allows for better training of deeper layers and earlier attention to the target anatomical structures of interest. This architecture offers up to a 7% relative improvement compared to the U-Net architecture in localization and detection performance, with an increase of only 0.75% in the number of parameters. Its performance is on par with, or slightly better than, the more computationally expensive U -Net++, which has 20% more parameters; making the proposed architecture a more computationally efficient alternative for real-time object detection and localization in ultrasound scans.
引用
收藏
页数:10
相关论文
共 55 条
  • [1] Al-Battal A.F., 2021, 2021 43 ANN INT C IE
  • [2] OBJECT DETECTION AND TRACKING IN ULTRASOUND SCANS USING AN OPTICAL FLOW AND SEMANTIC SEGMENTATION FRAMEWORK BASED ON CONVOLUTIONAL NEURAL NETWORKS
    Al-Battal, Abdullah F.
    Lerman, Imanuel R.
    Nguyen, Truong Q.
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 1096 - 1100
  • [3] Dataset of breast ultrasound images
    Al-Dhabyani, Walid
    Gomaa, Mohammed
    Khaled, Hussien
    Fahmy, Aly
    [J]. DATA IN BRIEF, 2020, 28
  • [4] Almekkawy M.K., 2014, Proceedings of Meetings on Acoustics 168ASA, V22
  • [5] Recurrent residual U-Net for medical image segmentation
    Alom, Md Zahangir
    Yakopcic, Chris
    Hasan, Mahmudul
    Taha, Tarek M.
    Asari, Vijayan K.
    [J]. JOURNAL OF MEDICAL IMAGING, 2019, 6 (01)
  • [6] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [7] SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound
    Baumgartner, Christian F.
    Kamnitsas, Konstantinos
    Matthew, Jacqueline
    Fletcher, Tara P.
    Smith, Sandra
    Koch, Lisa M.
    Kainz, Bernhard
    Rueckert, Daniel
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2017, 36 (11) : 2204 - 2215
  • [8] Bharadwaj S., 2020, IEEE INT ULTRA SYM, P1, DOI DOI 10.1109/ius46767.2020.9251739
  • [9] Bharadwaj S, 2020, IEEE ENG MED BIO, P2142, DOI 10.1109/EMBC44109.2020.9176269
  • [10] Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934