Combating Ambiguity for Hash-Code Learning in Medical Instance Retrieval

被引:4
作者
Fang, Jiansheng [1 ,2 ,3 ]
Fu, Huazhu [4 ]
Zeng, Dan [2 ]
Yan, Xiao
Yan, Yuguang [5 ]
Liu, Jiang [2 ,6 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Techol, Harbin 150001, Peoples R China
[2] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
[3] CVTE Res, Guangzhou 510530, Peoples R China
[4] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
[5] Univ Hong Kong, Dept Math, Hong Kong, Peoples R China
[6] Chinese Acad Sci, Cixi Inst Biomed Engn, Ningbo 315201, Peoples R China
关键词
Convolutional codes; Medical diagnostic imaging; Feature extraction; Visualization; Training; Semantics; Image segmentation; Medical instance retrieval; convolutional features; deep hashing methods; content-based image retrieval; CHEST RADIOGRAPHS; FEATURES; IMAGES;
D O I
10.1109/JBHI.2021.3082531
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
When encountering a dubious diagnostic case, medical instance retrieval can help radiologists make evidence-based diagnoses by finding images containing instances similar to a query case from a large image database. The similarity between the query case and retrieved similar cases is determined by visual features extracted from pathologically abnormal regions. However, the manifestation of these regions often lacks specificity, i.e., different diseases can have the same manifestation, and different manifestations may occur at different stages of the same disease. To combat the manifestation ambiguity in medical instance retrieval, we propose a novel deep framework called Y-Net, encoding images into compact hash-codes generated from convolutional features by feature aggregation. Y-Net can learn highly discriminative convolutional features by unifying the pixel-wise segmentation loss and classification loss. The segmentation loss allows exploring subtle spatial differences for good spatial-discriminability while the classification loss utilizes class-aware semantic information for good semantic-separability. As a result, Y-Net can enhance the visual features in pathologically abnormal regions and suppress the disturbing of the background during model training, which could effectively embed discriminative features into the hash-codes in the retrieval stage. Extensive experiments on two medical image datasets demonstrate that Y-Net can alleviate the ambiguity of pathologically abnormal regions and its retrieval performance outperforms the state-of-the-art method by an average of 9.27% on the returned list of 10.
引用
收藏
页码:3943 / 3954
页数:12
相关论文
共 66 条
[1]  
[Anonymous], 2016, ARXIV160606811
[2]  
[Anonymous], 2009, Advances in neural information processing systems
[3]  
Arandjelovic R, 2011, IEEE I CONF COMP VIS, P375, DOI 10.1109/ICCV.2011.6126265
[4]   Aggregating Deep Convolutional Features for Image Retrieval [J].
Babenko, Artem ;
Lempitsky, Victor .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1269-1277
[5]   Neural Codes for Image Retrieval [J].
Babenko, Artem ;
Slesarev, Anton ;
Chigorin, Alexandr ;
Lempitsky, Victor .
COMPUTER VISION - ECCV 2014, PT I, 2014, 8689 :584-599
[6]   SURF: Speeded up robust features [J].
Bay, Herbert ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION - ECCV 2006 , PT 1, PROCEEDINGS, 2006, 3951 :404-417
[7]  
Brownlee J., 2017, MACH LEARN MASTERY, V21
[8]   Order-Sensitive Deep Hashing for Multimorbidity Medical Image Retrieval [J].
Chen, Zhixiang ;
Cai, Ruojin ;
Lu, Jiwen ;
Feng, Jianjiang ;
Zhou, Jie .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2018, PT I, 2018, 11070 :620-628
[9]   Similarity regularized sparse group lasso for cup to disc ratio computation [J].
Cheng, Jun ;
Zhang, Zhuo ;
Tao, Dacheng ;
Wong, Damon Wing Kee ;
Liu, Jiang ;
Baskaran, Mani ;
Aung, Tin ;
Wong, Tien Yin .
BIOMEDICAL OPTICS EXPRESS, 2017, 8 (08) :3763-3777
[10]  
Conjeti Sailesh, 2017, Medical Image Computing and Computer Assisted Intervention MICCAI 2017. 20th International Conference. Proceedings: LNCS 10435, P550, DOI 10.1007/978-3-319-66179-7_63