Contrastive learning-based Adenoid Hypertrophy Grading Network Using Nasoendoscopic Image

被引:2
作者
Zheng, Siting [1 ,2 ]
Li, Xuechen [2 ,3 ]
Bi, Mingmin [4 ]
Wang, Yuxuan [1 ,2 ]
Liu, Haiyan [4 ]
Feng, Xiaoshan [4 ]
Fan, Yunping [4 ]
Shen, Linlin [1 ,2 ]
机构
[1] Shenzhen Univ, Coll Comp Sci & Software Engn, Shenzhen, Peoples R China
[2] Shenzhen Univ, AI Res Ctr Med Image Anal & Diag, Shenzhen, Peoples R China
[3] Shenzhen Univ, Natl Engn Lab Big Data Syst Comp Technol, Shenzhen, Peoples R China
[4] Sun Yat Sen Univ, Dept Otolaryngol, Affiliated Hosp 7, Shenzhen, Peoples R China
来源
2022 IEEE 35TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS) | 2022年
基金
中国国家自然科学基金;
关键词
adenoid hypertrophy; contrastive learning; deep learning; image classification; convolutional neural network;
D O I
10.1109/CBMS55023.2022.00074
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Adenoid hypertrophy is a common disease in children with otolaryngology diseases. Otolaryngologists usually use nasoendoscopy for adenoid hypertrophy screening, which is however tedious and time-consuming for the grading. So far, artificial intelligence technology has not been applied to the grading of nasoendoscopic adenoid. In this work, we firstly propose a novel multi-scale grading network, MIB-ANet, for adenoid hypertrophy classification. And we further propose a contrastive learning-based network to alleviate the overfitting problem of the model caused by lacking of nasoendoscopic adenoid images with high-quality annotations. The experimental results show that MIB-ANet shows the best grading performance compared to four classic CNNs, i.e., AlexNet, VGG16, ResNet50 and GoogleNet. Take F-1 score as an example, MIB-ANet achieves 1.38% higher F-1 score than the best baseline CNN - AlexNet. Due to the capability of the contrastive learning-based pre-training strategy in exploring unannotated data, the pre-training using SimCLR pretext task can consistently improve the performance of MIB-ANet when different ratios of the labeled training data are employed. The MIB-ANet pre-trained by SimCLR pretext task achieves 4.41%, 2.64%, 3.10%, and 1.71% higher F1 score when 25%, 50%, 75% and 100% of the training data are labeled, respectively.
引用
收藏
页码:377 / 382
页数:6
相关论文
共 23 条
[1]  
Albashish Dheeb, 2021, 2021 International Conference on Information Technology (ICIT), P805, DOI 10.1109/ICIT52682.2021.9491631
[2]  
Ali Y., 2021, arXiv
[3]   GoogLeNet-Based Ensemble FCNet Classifier for Focal Liver Lesion Diagnosis [J].
Balagourouchetty, Lakshmipriya ;
Pragatheeswaran, Jayanthi K. ;
Pottakkat, Biju ;
Ramkumar, G. .
IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2020, 24 (06) :1686-1694
[4]  
Chen T, 2020, PR MACH LEARN RES, V119
[5]   TW-GAN: Topology and width aware GAN for retinal artery/vein classification [J].
Chen, Wenting ;
Yu, Shuang ;
Ma, Kai ;
Ji, Wei ;
Bian, Cheng ;
Chu, Chunyan ;
Shen, Linlin ;
Zheng, Yefeng .
MEDICAL IMAGE ANALYSIS, 2022, 77
[6]  
Grill, 2020, arXiv
[7]   Fully Convolutional Boundary Regression for Retina OCT Segmentation [J].
He, Yufan ;
Carass, Aaron ;
Liu, Yihao ;
Jedynak, Bruno M. ;
Solomon, Sharon D. ;
Saidha, Shiv ;
Calabresi, Peter A. ;
Prince, Jerry L. .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT I, 2019, 11764 :120-128
[8]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[9]   TRIBYOL: TRIPLET BYOL FOR SELF-SUPERVISED REPRESENTATION LEARNING [J].
Li, Guang ;
Togo, Ren ;
Ogawa, Takahiro ;
Haseyama, Miki .
2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, :3458-3462
[10]   Multi-resolution convolutional networks for chest X-ray radiograph based lung nodule detection [J].
Li, Xuechen ;
Shen, Linlin ;
Xie, Xinpeng ;
Huang, Shiyun ;
Xie, Zhien ;
Hong, Xian ;
Yu, Juan .
ARTIFICIAL INTELLIGENCE IN MEDICINE, 2020, 103