Masked autoencoders with handcrafted feature predictions: Transformer for weakly supervised esophageal cancer classification

被引:4
作者
Bai, Yunhao [1 ]
Li, Wenqi [2 ]
An, Jianpeng [1 ]
Xia, Lili [1 ]
Chen, Huazhen [1 ]
Zhao, Gang [2 ]
Gao, Zhongke [1 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin, Peoples R China
[2] Tianjin Med Univ, Natl Clin Res Ctr Canc, Dept Pathol, Key Lab Canc Prevent & Therapy,Canc Inst & Hosp, Tianjin, Peoples R China
基金
中国国家自然科学基金;
关键词
Self-supervised learning; Esophageal cancer; Multi-instance learning; Whole slide image analysis; Transformer; WHOLE-SLIDE IMAGES;
D O I
10.1016/j.cmpb.2023.107936
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background and Objective: Esophageal cancer is a serious disease with a high prevalence in Eastern Asia. Histopathology tissue analysis stands as the gold standard in diagnosing esophageal cancer. In recent years, there has been a shift towards digitizing histopathological images into whole slide images (WSIs), progressively integrating them into cancer diagnostics. However, the gigapixel sizes of WSIs present significant storage and processing challenges, and they often lack localized annotations. To address this issue, multi-instance learning (MIL) has been introduced for WSI classification, utilizing weakly supervised learning for diagnosis analysis. By applying the principles of MIL to WSI analysis, it is possible to reduce the workload of pathologists by facilitating the generation of localized annotations. Nevertheless, the approach's effectiveness is hindered by the traditional simple aggregation operation and the domain shift resulting from the prevalent use of convolutional feature extractors pretrained on ImageNet.Methods: We propose a MIL-based framework for WSI analysis and cancer classification. Concurrently, we introduce employing self-supervised learning, which obviates the need for manual annotation and demonstrates versatility in various tasks, to pretrain feature extractors. This method enhances the extraction of representative features from esophageal WSI for MIL, ensuring more robust and accurate performance.Results: We build a comprehensive dataset of whole esophageal slide images and conduct extensive experiments utilizing this dataset. The performance on our dataset demonstrates the efficiency of our proposed MIL framework and the pretraining process, with our framework outperforming existing methods, achieving an accuracy of 93.07% and AUC (area under the curve) of 95.31%.Conclusion: This work proposes an effective MIL method to classify WSI of esophageal cancer. The promising results indicate that our cancer classification framework holds great potential in promoting the automatic whole esophageal slide image analysis.
引用
收藏
页数:12
相关论文
共 64 条
[21]   Multi-scale Domain-adversarial Multiple-instance CNN for Cancer Subtype Classification with Unannotated Histopathological Images [J].
Hashimoto, Noriaki ;
Fukushima, Daisuke ;
Koga, Ryoichi ;
Takagi, Yusuke ;
Ko, Kaho ;
Kohno, Kei ;
Nakaguro, Masato ;
Nakamura, Shigeo ;
Hontani, Hidekata ;
Takeuchi, Ichiro .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3851-3860
[22]   Masked Autoencoders Are Scalable Vision Learners [J].
He, Kaiming ;
Chen, Xinlei ;
Xie, Saining ;
Li, Yanghao ;
Dollar, Piotr ;
Girshick, Ross .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :15979-15988
[23]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[24]   Patch-based Convolutional Neural Network for Whole Slide Tissue Image Classification [J].
Hou, Le ;
Samaras, Dimitris ;
Kurc, Tahsin M. ;
Gao, Yi ;
Davis, James E. ;
Saltz, Joel H. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2424-2433
[25]   Integration of Patch Features Through Self-supervised Learning and Transformer for Survival Analysis on Whole Slide Images [J].
Huang, Ziwang ;
Chai, Hua ;
Wang, Ruoqi ;
Wang, Haitao ;
Yang, Yuedong ;
Wu, Hejun .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VIII, 2021, 12908 :561-570
[26]  
Ikromjanov K., 2022, IEEE INT C ART INT I, P399, DOI DOI 10.1109/ICAIIC54071.2022.9722635
[27]  
Ilse M, 2018, PR MACH LEARN RES, V80
[28]  
Jiang YF, 2021, ADV NEUR IN
[29]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[30]   Computer-aided diagnosis from weak supervision: A benchmarking study [J].
Kandemir, Melih ;
Hamprecht, Fred A. .
COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2015, 42 :44-50