A Transformer-Based Knowledge Distillation Network for Cortical Cataract Grading

被引:1
|
作者
Wang, Jinhong [1 ,2 ]
Xu, Zhe [3 ]
Zheng, Wenhao [1 ,2 ]
Ying, Haochao [4 ]
Chen, Tingting [1 ,2 ]
Liu, Zuozhu [5 ]
Chen, Danny Z. [6 ]
Yao, Ke [3 ]
Wu, Jian [7 ,8 ]
机构
[1] Zhejiang Univ, Affiliated Hosp 2, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[2] Zhejiang Univ, Affiliated Hosp 2, Eye Ctr, Hangzhou 310027, Peoples R China
[3] Zhejiang Univ, Affiliated Hosp 2, Eye Ctr, Sch Med, Hangzhou 310009, Zhejiang, Peoples R China
[4] Zhejiang Univ, Sch Publ Hlth, Hangzhou 310058, Peoples R China
[5] Zhejiang Univ, ZJU UIUC Inst, Res & Dev Ctr Intelligent Healthcare, ZJU Angelalign Inc, Haining 310058, Peoples R China
[6] Univ Notre Dame, Dept Comp Sci & Engn, Notre Dame, IN 46556 USA
[7] Zhejiang Univ, Affiliated Hosp 2, Sch Med, Sch Publ Hlth, Hangzhou 310058, Peoples R China
[8] Zhejiang Univ, Inst Wenzhou, Hangzhou 310058, Peoples R China
关键词
Cataracts; Transformers; Annotations; Feature extraction; Image edge detection; Fuses; Knowledge engineering; Cataract grading; knowledge distillation; transformer; medical imaging classification; CLASSIFICATION; IMAGES;
D O I
10.1109/TMI.2023.3327274
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Cortical cataract, a common type of cataract, is particularly difficult to be diagnosed automatically due to the complex features of the lesions. Recently, many methods based on edge detection or deep learning were proposed for automatic cataract grading. However, these methods suffer a large performance drop in cortical cataract grading due to the more complex cortical opacities and uncertain data. In this paper, we propose a novel Transformer-based Knowledge Distillation Network, called TKD-Net, for cortical cataract grading. To tackle the complex opacity problem, we first devise a zone decomposition strategy to extract more refined features and introduce special sub-scores to consider critical factors of clinical cortical opacity assessment (location, area, density) for comprehensive quantification. Next, we develop a multi-modal mix-attention Transformer to efficiently fuse sub-scores and image modality for complex feature learning. However, obtaining the sub-score modality is a challenge in the clinic, which could cause the modality missing problem instead. To simultaneously alleviate the issues of modality missing and uncertain data, we further design a Transformer-based knowledge distillation method, which uses a teacher model with perfect data to guide a student model with modality-missing and uncertain data. We conduct extensive experiments on a dataset of commonly-used slit-lamp images annotated by the LOCS III grading system to demonstrate that our TKD-Net outperforms state-of-the-art methods, as well as the effectiveness of its key components.
引用
收藏
页码:1089 / 1101
页数:13
相关论文
共 50 条
  • [41] A transformer based visual tracker with restricted token interaction and knowledge distillation
    Liu, Nian
    Zhang, Yi
    KNOWLEDGE-BASED SYSTEMS, 2025, 307
  • [42] A Lightweight Object Counting Network Based on Density Map Knowledge Distillation
    Shen, Zhilong
    Li, Guoquan
    Xia, Ruiyang
    Meng, Hongying
    Huang, Zhengwen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1492 - 1505
  • [43] Transformer-based attention network for stock movement prediction
    Zhang, Qiuyue
    Qin, Chao
    Zhang, Yunfeng
    Bao, Fangxun
    Zhang, Caiming
    Liu, Peide
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 202
  • [44] Transformer-Based Light Field Salient Object Detection and Its Application to Autofocus
    Jiang, Yao
    Li, Xin
    Fu, Keren
    Zhao, Qijun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 6647 - 6659
  • [45] A Novel Transformer-Based Attention Network for Image Dehazing
    Gao, Guanlei
    Cao, Jie
    Bao, Chun
    Hao, Qun
    Ma, Aoqi
    Li, Gang
    SENSORS, 2022, 22 (09)
  • [46] A Transformer-Based Network for Deformable Medical Image Registration
    Wang, Yibo
    Qian, Wen
    Li, Mengqi
    Zhang, Xuming
    ARTIFICIAL INTELLIGENCE, CICAI 2022, PT I, 2022, 13604 : 502 - 513
  • [47] A Transformer-Based Model With Self-Distillation for Multimodal Emotion Recognition in Conversations
    Ma, Hui
    Wang, Jian
    Lin, Hongfei
    Zhang, Bo
    Zhang, Yijia
    Xu, Bo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 776 - 788
  • [48] Knowledge Distillation for Streaming Transformer-Transducer
    Kojima, Atsushi
    INTERSPEECH 2021, 2021, : 2841 - 2845
  • [49] Transformer-based ASR Incorporating Time-reduction Layer and Fine-tuning with Self-Knowledge Distillation
    Haidar, Md Akmal
    Xing, Chao
    Rezagholizadeh, Mehdi
    INTERSPEECH 2021, 2021, : 2102 - 2106
  • [50] TMA-Net: A Transformer-Based Multi-Scale Attention Network for Surgical Instrument Segmentation
    Yang, Lei
    Wang, Hongyong
    Gu, Yuge
    Bian, Guibin
    Liu, Yanhong
    Yu, Hongnian
    IEEE TRANSACTIONS ON MEDICAL ROBOTICS AND BIONICS, 2023, 5 (02): : 323 - 334