Consistency Loss for Improved Colonoscopy Landmark Detection with Vision Transformers

被引:0
作者
Tamhane, Aniruddha [1 ]
Dobkin, Daniel [1 ]
Shtalrid, Ore [1 ]
Bouhnik, Moshe [1 ]
Posner, Erez [1 ]
Mida, Tse'ela [1 ]
机构
[1] Intuit Surg Inc, 1020 Kifer Rd, Sunnyvale, CA 94086 USA
来源
MACHINE LEARNING IN MEDICAL IMAGING, MLMI 2023, PT II | 2024年 / 14349卷
关键词
Colonoscopy; Vision Transformer; Landmark Detection; Self-supervised learning; Consistency loss; Data sampling; COLON;
D O I
10.1007/978-3-031-45676-3_13
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Colonoscopy is a procedure used to examine the colon and rectum for colorectal cancer or other abnormalities including polyps or diverticula. Apart from the actual diagnosis, manually processing the snapshots taken during the colonoscopy procedure (for medical record keeping) consumes a large amount of the clinician's time. This can be automated through post-procedural machine learning based algorithms which classify anatomical landmarks in the colon. In this work, we have developed a pipeline for training vision-transformers for identifying anatomical landmarks, including appendiceal orifice, ileocecal valve/cecum landmark and rectum retroflection. To increase the accuracy of the model, we utilize a hybrid approach that combines algorithm-level and data-level techniques. We introduce a consistency loss to enhance model immunity to label inconsistencies, as well as a semantic non-landmark sampling technique aimed at increasing focus on colonic findings. For training and testing our pipeline, we have annotated 307 colonoscopy videos and 2363 snapshots with the assistance of several medical experts for enhanced reliability. The algorithm identifies landmarks with an accuracy of 92% on the test dataset.
引用
收藏
页码:124 / 133
页数:10
相关论文
共 50 条
  • [21] Revolutionizing dementia detection: Leveraging vision and Swin transformers for early diagnosis
    Rini, P. L.
    Gayathri, K. S.
    AMERICAN JOURNAL OF MEDICAL GENETICS PART B-NEUROPSYCHIATRIC GENETICS, 2024, 195 (07)
  • [22] ViP: Unified Certified Detection and Recovery for Patch Attack with Vision Transformers
    Li, Junbo
    Zhang, Huan
    Xie, Cihang
    COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 573 - 587
  • [23] Artificial Cognition for Early Leaf Disease Detection using Vision Transformers
    Huy-Tan Thai
    Nhu-Y Tran-Van
    Kim-Hung Le
    2021 INTERNATIONAL CONFERENCE ON ADVANCED TECHNOLOGIES FOR COMMUNICATIONS (ATC 2021), 2021, : 33 - 38
  • [24] Towards improved fundus disease detection using Swin Transformers
    Jawad, M. Abdul
    Khursheed, Farida
    Nawaz, Shah
    Mir, A. H.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (32) : 78125 - 78159
  • [25] Central loss guides coordinated Transformer for reliable anatomical landmark detection
    Zhu, Qikui
    Bi, Yihui
    Chen, Jie
    Chu, Xiangpeng
    Wang, Danxin
    Wang, Yanqing
    NEURAL NETWORKS, 2025, 187
  • [26] Improved Colon Navigation for Efficient Polyp Detection in Virtual Colonoscopy
    Ismail, Marwa
    Farag, Aly
    Hassouna, M. Sabry
    Dryden, Gerald
    Falk, Robert
    ABDOMINAL IMAGING: COMPUTATION AND CLINICAL APPLICATIONS, 2013, 8198 : 117 - 125
  • [27] Combining Vision Transformers and crane load information for a rope winding detection system
    Davide Picchi
    Sigrid Brell-Cokcan
    Construction Robotics, 2025, 9 (1)
  • [28] Learning From Synthetic InSAR With Vision Transformers: The Case of Volcanic Unrest Detection
    Bountos, Nikolaos Ioannis
    Michail, Dimitrios
    Papoutsis, Ioannis
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [29] Deep Neural Networks and Computer Vision for Real-Time Facial Landmark Detection
    Ciobotaru, Alexandru
    Gota, Dan Ioan
    Puscasiu, Adela
    Stan, Ovidiu
    Fanca, Alexandra
    Domnuta, Claudiu
    Valean, Honoriu
    Miclea, Liviu
    INTERNATIONAL CONFERENCE ON ELECTRICAL, COMPUTER AND ENERGY TECHNOLOGIES (ICECET 2021), 2021, : 424 - 429
  • [30] DeepFake detection algorithm based on improved vision transformer
    Heo, Young-Jin
    Yeo, Woon-Ha
    Kim, Byung-Gyu
    APPLIED INTELLIGENCE, 2023, 53 (07) : 7512 - 7527