Image Compositing for Segmentation of Surgical Tools Without Manual Annotations

被引:29
作者
Garcia-Peraza-Herrera, Luis C. [1 ,2 ]
Fidon, Lucas [2 ]
D'Ettorre, Claudia [3 ]
Stoyanov, Danail [3 ]
Vercauteren, Tom [2 ]
Ourselin, Sebastien [2 ]
机构
[1] UCL, Dept Med Phys & Biomed Engn, London WC1E 6BT, England
[2] Kings Coll London, Dept Surg & Intervent Engn, London WC2R 2LS, England
[3] UCL, Dept Comp Sci, London WC1E 6BT, England
基金
英国工程与自然科学研究理事会; 欧盟地平线“2020”;
关键词
Image segmentation; Instruments; Tools; Training; Task analysis; Surgery; Manuals; Image compositing; chroma key; tool segmentation;
D O I
10.1109/TMI.2021.3057884
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Producing manual, pixel-accurate, image segmentation labels is tedious and time-consuming. This is often a rate-limiting factor when large amounts of labeled images are required, such as for training deep convolutional networks for instrument-background segmentation in surgical scenes. No large datasets comparable to industry standards in the computer vision community are available for this task. To circumvent this problem, we propose to automate the creation of a realistic training dataset by exploiting techniques stemming from special effects and harnessing them to target training performance rather than visual appeal. Foreground data is captured by placing sample surgical instruments over a chroma key (a.k.a. green screen) in a controlled environment, thereby making extraction of the relevant image segment straightforward. Multiple lighting conditions and viewpoints can be captured and introduced in the simulation by moving the instruments and camera and modulating the light source. Background data is captured by collecting videos that do not contain instruments. In the absence of pre-existing instrument-free background videos, minimal labeling effort is required, just to select frames that do not contain surgical instruments from videos of surgical interventions freely available online. We compare different methods to blend instruments over tissue and propose a novel data augmentation approach that takes advantage of the plurality of options. We show that by training a vanilla U-Net on semi-synthetic data only and applying a simple post-processing, we are able to match the results of the same network trained on a publicly available manually labeled real dataset.
引用
收藏
页码:1450 / 1460
页数:11
相关论文
共 41 条
  • [1] Allan M., 2019, 190206426 ARXIV
  • [2] [Anonymous], 2017, P 31 INT C NEURAL IN, DOI DOI 10.5555/3295222.3295408
  • [3] [Anonymous], 2017, CORR
  • [4] A MULTIRESOLUTION SPLINE WITH APPLICATION TO IMAGE MOSAICS
    BURT, PJ
    ADELSON, EH
    [J]. ACM TRANSACTIONS ON GRAPHICS, 1983, 2 (04): : 217 - 236
  • [5] Albumentations: Fast and Flexible Image Augmentations
    Buslaev, Alexander
    Iglovikov, Vladimir I.
    Khvedchenya, Eugene
    Parinov, Alex
    Druzhinin, Mikhail
    Kalinin, Alexandr A.
    [J]. INFORMATION, 2020, 11 (02)
  • [6] Synthesizing Training Images for Boosting Human 3D Pose Estimation
    Chen, Wenzheng
    Wang, Huan
    Li, Yangyan
    Su, Hao
    Wang, Zhenhua
    Tu, Changhe
    Lischinski, Dani
    Cohen-Or, Daniel
    Chen, Baoquan
    [J]. PROCEEDINGS OF 2016 FOURTH INTERNATIONAL CONFERENCE ON 3D VISION (3DV), 2016, : 479 - 488
  • [7] Modeling Visual Context Is Key to Augmenting Object Detection Datasets
    Dvornik, Nikita
    Mairal, Julien
    Schmid, Cordelia
    [J]. COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 : 375 - 391
  • [8] Cut, Paste and Learn: Surprisingly Easy Synthesis for Instance Detection
    Dwibedi, Debidatta
    Misra, Ishan
    Hebert, Martial
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1310 - 1319
  • [9] Real-Time Segmentation of Non-rigid Surgical Tools Based on Deep Learning and Tracking
    Garcia-Peraza-Herrera, Luis C.
    Li, Wenqi
    Gruijthuijsen, Caspar
    Devreker, Alain
    Attilakos, George
    Deprest, Jan
    Vander Poorten, Emmanuel
    Stoyanov, Danail
    Vercauteren, Tom
    Ourselin, Sebastien
    [J]. COMPUTER-ASSISTED AND ROBOTIC ENDOSCOPY, 2017, 10170 : 84 - 95
  • [10] Haase S, 2013, IEEE WORK APP COMP, P449, DOI 10.1109/WACV.2013.6475053