Uncertainty-aware prototypical learning for anomaly detection in medical images

被引:5
作者
Huang, Chao [1 ,2 ]
Shi, Yushu [2 ]
Zhang, Bob [1 ]
Lyu, Ke [3 ,4 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, PAMI Res Grp, Taipa 519000, Peoples R China
[2] Sun Yat Sen Univ, Sch Cyber Sci & Technol, Shenzhen Campus, Shenzhen 518107, Peoples R China
[3] Univ Chinese Acad Sci, Sch Engn Sci, Beijing 100049, Peoples R China
[4] Pengcheng Lab, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Anomalous object detection; Medical image analysis; Prototypical learning;
D O I
10.1016/j.neunet.2024.106284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Anomalous object detection (AOD) in medical images aims to recognize the anomalous lesions, and is crucial for early clinical diagnosis of various cancers. However, it is a difficult task because of two reasons: (1) the diversity of the anomalous lesions and (2) the ambiguity of the boundary between anomalous lesions and their normal surroundings. Unlike existing single -modality AOD models based on deterministic mapping, we constructed a probabilistic and deterministic AOD model. Specifically, we designed an uncertaintyaware prototype learning framework, which considers the diversity and ambiguity of anomalous lesions. A prototypical learning transformer (Pformer) is established to extract and store the prototype features of different anomalous lesions. Moreover, Bayesian neural uncertainty quantizer, a probabilistic model, is designed to model the distributions over the outputs of the model to measure the uncertainty of the model's detection results for each pixel. Essentially, the uncertainty of the model's anomaly detection result for a pixel can reflect the anomalous ambiguity of this pixel. Furthermore, an uncertainty -guided reasoning transformer (Uformer) is devised to employ the anomalous ambiguity, encouraging the proposed model to focus on pixels with high uncertainty. Notably, prototypical representations stored in Pformer are also utilized in anomaly reasoning that enables the model to perceive diversities of the anomalous objects. Extensive experiments on five benchmark datasets demonstrate the superiority of our proposed method. The source code will be available in github.com/umchaohuang/UPformer.
引用
收藏
页数:10
相关论文
共 47 条
[41]   Deep Model Intellectual Property Protection via Deep Watermarking [J].
Zhang, Jie ;
Chen, Dongdong ;
Liao, Jing ;
Zhang, Weiming ;
Feng, Huamin ;
Hua, Gang ;
Yu, Nenghai .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) :4005-4020
[42]   SwIPE: Efficient and Robust Medical Image Segmentation with Implicit Patch Embeddings [J].
Zhang, Yejia ;
Gu, Pengfei ;
Sapkota, Nishchal ;
Chen, Danny Z. .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT V, 2023, 14224 :315-326
[43]   Deep learning in food category recognition [J].
Zhang, Yudong ;
Deng, Lijia ;
Zhu, Hengde ;
Wang, Wei ;
Ren, Zeyu ;
Zhou, Qinghua ;
Lu, Siyuan ;
Sun, Shiting ;
Zhu, Ziquan ;
Gorriz, Juan Manuel ;
Wang, Shuihua .
INFORMATION FUSION, 2023, 98
[44]   Road Extraction by Deep Residual U-Net [J].
Zhang, Zhengxin ;
Liu, Qingjie ;
Wang, Yunhong .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2018, 15 (05) :749-753
[45]   Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers [J].
Zheng, Sixiao ;
Lu, Jiachen ;
Zhao, Hengshuang ;
Zhu, Xiatian ;
Luo, Zekun ;
Wang, Yabiao ;
Fu, Yanwei ;
Feng, Jianfeng ;
Xiang, Tao ;
Torr, Philip H. S. ;
Zhang, Li .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :6877-6886
[46]   UNet plus plus : A Nested U-Net Architecture for Medical Image Segmentation [J].
Zhou, Zongwei ;
Siddiquee, Md Mahfuzur Rahman ;
Tajbakhsh, Nima ;
Liang, Jianming .
DEEP LEARNING IN MEDICAL IMAGE ANALYSIS AND MULTIMODAL LEARNING FOR CLINICAL DECISION SUPPORT, DLMIA 2018, 2018, 11045 :3-11
[47]  
登范, 2021, Scientia Sinica Informationis, V51, P1475, DOI [10.1360/ssi-2020-0370, DOI 10.1360/SSI-2020-0370, 10.1360/SSI-2020-0370]