Robotic Instrument Segmentation With Image-to-Image Translation

被引:21
|
作者
Colleoni, Emanuele [1 ]
Stoyanov, Danail [1 ]
机构
[1] Univ Coll London UCL, Wellcome EPSRC Ctr Intervent & Surg Sci WEISS, London W1W 7TS, England
基金
欧盟地平线“2020”; 英国工程与自然科学研究理事会;
关键词
Image segmentation; Gallium nitride; Feature extraction; Instruments; Robots; Generative adversarial networks; Data models; Medical robots and systems; deep learning methods; image-to-image translation; surgical robot simulators; surgical tool segmentation;
D O I
10.1109/LRA.2021.3056354
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
The semantic segmentation of robotic surgery video and the delineation of robotic instruments are important for enabling automation. Despite major recent progresses, the majority of the latest deep learning models for instrument detection and segmentation rely on large datasets with ground truth labels. While demonstrating the capability, reliance on large labelled data is a problem for practical applications because systems would need to be re-trained on domain variations such as procedure type or instrument sets. In this letter, we propose to alleviate this problem by training deep learning models on datasets that are synthesised using image-to-image translation techniques and we investigate different methods to perform this process optimally. Experimentally, we demonstrate that the same deep network architecture for robotic instrument segmentation can be trained on both real data and on our proposed synthetic data without affecting the quality of the output models' performance. We show this for several recent approaches and provide experimental support on publicly available datasets, which highlight the potential value of this approach.
引用
收藏
页码:935 / 942
页数:8
相关论文
共 50 条
  • [41] CycleSAR: SAR Image Despeckling as Unpaired Image-to-Image Translation
    Lattari, Francesco
    Santomarco, Vincenzo
    Santambrogio, Riccardo
    Rucci, Alessio
    Matteucci, Matteo
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [42] Generating Adversarial Examples in One Shot With Image-to-Image Translation GAN
    Zhang, Weijia
    IEEE ACCESS, 2019, 7 : 151103 - 151119
  • [43] Multidomain image-to-image translation model based on hidden space sharing
    Ding Yuxin
    Wang Longfei
    Neural Computing and Applications, 2022, 34 : 283 - 298
  • [44] VecGAN: Image-to-Image Translation with Interpretable Latent Directions
    Dalva, Yusuf
    Altindis, Said Fahri
    Dundar, Aysegul
    COMPUTER VISION - ECCV 2022, PT XVI, 2022, 13676 : 153 - 169
  • [45] MULTIMODAL IMAGE-TO-IMAGE TRANSLATION FOR GENERATION OF GASTRITIS IMAGES
    Togo, Ren
    Ogawa, Takahiro
    Haseyama, Miki
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 2466 - 2470
  • [46] Multimodal Unsupervised Image-to-Image Translation Without Independent Style Encoder
    Sun, Yanbei
    Lu, Yao
    Lu, Haowei
    Zhao, Qingjie
    Wang, Shunzhou
    MULTIMEDIA MODELING (MMM 2022), PT I, 2022, 13141 : 624 - 636
  • [47] Unsupervised Multimodal Image-to-Image Translation: Generate What You Want
    Zhang, Chao
    Xi, Wei
    Liu, Xinhui
    Bai, Gairui
    Sun, Jingtong
    Yu, Fan
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [48] Multidomain image-to-image translation model based on hidden space sharing
    Ding, Yuxin
    Wang, Longfei
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (01) : 283 - 298
  • [49] Object Detection to Evaluate Image-to-Image Translation on Different Road Conditions
    Sudo, Fumiya
    Hashimoto, Yoshihiro
    Lisi, Giuseppe
    INTELLIGENT HUMAN SYSTEMS INTEGRATION 2020, 2020, 1131 : 143 - 149
  • [50] Continuous and Diverse Image-to-Image Translation via Signed Attribute Vectors
    Mao, Qi
    Tseng, Hung-Yu
    Lee, Hsin-Ying
    Huang, Jia-Bin
    Ma, Siwei
    Yang, Ming-Hsuan
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (02) : 517 - 549