Clustering-Guided Twin Contrastive Learning for Endomicroscopy Image Classification

被引:0
作者
Zhou, Jingjun [1 ]
Dong, Xiangjiang [2 ]
Liu, Qian [1 ,3 ]
机构
[1] Hainan Univ, Sch Biomed Engn, Haikou 570228, Peoples R China
[2] Huazhong Univ Sci & Technol, Wuhan Natl Lab Optoelect, Wuhan 430074, Peoples R China
[3] Hainan Univ, Sch Biomed Engn, Key Lab Biomed Engn Hainan Prov, Haikou 570228, Peoples R China
基金
中国国家自然科学基金;
关键词
Clustering; contrastive learning; image classification and gastrointestinal; probe-based confocal laser endomicroscopy (pCLE); CONFOCAL LASER ENDOMICROSCOPY; SAFETY;
D O I
10.1109/JBHI.2024.3366223
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Learning better representations is essential in medical image analysis for computer-aided diagnosis. However, learning discriminative semantic features is a major challenge due to the lack of large-scale well-annotated datasets. Thus, how can we learn a well-structured categorizable embedding space in limited-scale and unlabeled datasets? In this paper, we proposed a novel clustering-guided twin-contrastive learning framework (CTCL) that learns the discriminative representations of probe-based confocal laser endomicroscopy (pCLE) images for gastrointestinal (GI) tumor classification. Compared with traditional contrastive learning, in which only two randomly augmented views of the same instance are considered, the proposed CTCL aligns more semantically related and class-consistent samples by clustering, which improved intra-class tightness and inter-class variability to produce more informative representations. Furthermore, based on the inherent properties of CLE (geometric invariance and intrinsic noise), we proposed to regard CLE images with any angle rotation and CLE images with different noises as the same instance, respectively, for increased variability and diversity of samples. By optimizing CTCL in an end-to-end expectation-maximization framework, comprehensive experimental results demonstrated that CTCL-based visual representations achieved competitive performance on each downstream task as well as more robustness and transferability compared with existing state-of-the-art SSL and supervised methods. Notably, CTCL achieved 75.60%/78.45% and 64.12%/77.37% top-1 accuracy on the linear evaluation protocol and few-shot classification downstream tasks, respectively, which outperformed the previous best results by 1.27%/1.63% and 0.5%/3%, respectively. The proposed method holds great potential to assist pathologists in achieving an automated, fast, and high-precision diagnosis of GI tumors and accurately determining different stages of tumor development based on CLE images.
引用
收藏
页码:2879 / 2890
页数:12
相关论文
共 56 条
  • [31] Self-Path: Self-Supervision for Classification of Pathology Images With Limited Annotations
    Koohbanani, Navid Alemi
    Unnikrishnan, Balagopal
    Khurram, Syed Ali
    Krishnaswamy, Pavitra
    Rajpoot, Nasir
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2021, 40 (10) : 2845 - 2856
  • [32] Learning Representations for Automatic Colorization
    Larsson, Gustav
    Maire, Michael
    Shakhnarovich, Gregory
    [J]. COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 : 577 - 593
  • [33] Li P., 2021, P INT C LEARN REPR
  • [34] Revisiting Local Descriptor based Image-to-Class Measure for Few-shot Learning
    Li, Wenbin
    Wang, Lei
    Xu, Jinglin
    Huo, Jing
    Gao, Yang
    Luo, Jiebo
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7253 - 7260
  • [35] Twin Contrastive Learning for Online Clustering
    Li, Yunfan
    Yang, Mouxing
    Peng, Dezhong
    Li, Taihao
    Huang, Jiantao
    Peng, Xi
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (09) : 2205 - 2221
  • [36] A survey on deep learning in medical image analysis
    Litjens, Geert
    Kooi, Thijs
    Bejnordi, Babak Ehteshami
    Setio, Arnaud Arindra Adiyoso
    Ciompi, Francesco
    Ghafoorian, Mohsen
    van der Laak, Jeroen A. W. M.
    van Ginneken, Bram
    Sanchez, Clara I.
    [J]. MEDICAL IMAGE ANALYSIS, 2017, 42 : 60 - 88
  • [37] Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles
    Noroozi, Mehdi
    Favaro, Paolo
    [J]. COMPUTER VISION - ECCV 2016, PT VI, 2016, 9910 : 69 - 84
  • [38] Self-Supervised Learning for Few-Shot Medical Image Segmentation
    Ouyang, Cheng
    Biffi, Carlo
    Chen, Chen
    Kart, Turkay
    Qiu, Huaqi
    Rueckert, Daniel
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2022, 41 (07) : 1837 - 1848
  • [39] Paszke A, 2019, ADV NEUR IN, V32
  • [40] Context Encoders: Feature Learning by Inpainting
    Pathak, Deepak
    Krahenbuhl, Philipp
    Donahue, Jeff
    Darrell, Trevor
    Efros, Alexei A.
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2536 - 2544