Deconv-transformer (DecT): A histopathological image classification model for breast cancer based on color deconvolution and transformer architecture

被引:108
作者
He, Zhu [1 ]
Lin, Mingwei [1 ]
Xu, Zeshui [3 ]
Yao, Zhiqiang [1 ]
Chen, Hong [2 ]
Alhudhaif, Adi [4 ]
Alenezi, Fayadh [5 ]
机构
[1] Fujian Normal Univ, Coll Comp & Cyber Secur, Fuzhou 350117, Fujian, Peoples R China
[2] Fujian Normal Univ, Sch Math & Stat, Fuzhou 350117, Fujian, Peoples R China
[3] Sichuan Univ, Business Sch, Chengdu 610064, Sichuan, Peoples R China
[4] Prince Sattam Bin Abdulaziz Univ, Coll Comp Engn & Sci Al Kharj, Dept Comp Sci, POB 151, Al Kharj 11942, Saudi Arabia
[5] Jouf Univ, Coll Engn, Dept Elect Engn, Sakaka, Saudi Arabia
关键词
Histopathological image; Breast cancer; Deep learning; Color deconvolution; Color space; RECOGNITION;
D O I
10.1016/j.ins.2022.06.091
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Histopathological image recognition of breast cancer is an onerous task. Although many deep learning models have achieved good classification results on histopathological image classification tasks, these models do not take full advantage of the staining properties of histopathological images. In this paper, we propose a novel Deconv-Transformer (DecT) network model, which incorporates the color deconvolution in the form of convolution layers. This model uses a self-attention mechanism to match the independent properties of the HED channel information obtained by the color deconvolution. It also uses a method similar to the residual connection to fuse the information of both RGB and HED color space images, which can compensate for the information loss in the process of transferring RGB images to HED images. The training process of the DecT model is divided into two stages so that the parameters of the deconvolution layer can be better adapted to different types of histopathological images. We use the color jitter in the image data augmentation process to reduce the overfitting in the model training process. The DecT model achieves an average accuracy of 93.02% and F1-score of 0.9389 on BreakHis dataset, and an average accuracy of 79.06% and 81.36% on BACH and UC datasets. (C) 2022 Elsevier Inc. All rights reserved.
引用
收藏
页码:1093 / 1112
页数:20
相关论文
共 47 条
[11]  
Dosovitskiy A, 2021, Arxiv, DOI [arXiv:2010.11929, DOI 10.48550/ARXIV.2010.11929]
[12]  
Elmore JG, 2015, JAMA-J AM MED ASSOC, V313, P1122, DOI 10.1001/jama.2015.1405
[13]   Multilingual handwritten numeral recognition using a robust deep network joint with transfer learning [J].
Fateh, Amirreza ;
Fateh, Mansoor ;
Abolghasemi, Vahid .
INFORMATION SCIENCES, 2021, 581 :479-494
[14]   Cancer statistics for the year 2020: An overview [J].
Ferlay, Jacques ;
Colombet, Murielle ;
Soerjomataram, Isabelle ;
Parkin, Donald M. ;
Pineros, Marion ;
Znaor, Ariana ;
Bray, Freddie .
INTERNATIONAL JOURNAL OF CANCER, 2021, 149 (04) :778-789
[15]   Instance-Based Vision Transformer for Subtyping of Papillary Renal Cell Carcinoma in Histopathological Image [J].
Gao, Zeyu ;
Hong, Bangyang ;
Zhang, Xianli ;
Li, Yang ;
Jia, Chang ;
Wu, Jialun ;
Wang, Chunbao ;
Meng, Deyu ;
Li, Chen .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2021, PT VIII, 2021, 12908 :299-308
[16]   LeViT: a Vision Transformer in ConvNet's Clothing for Faster Inference [J].
Graham, Ben ;
El-Nouby, Alaaeldin ;
Touvron, Hugo ;
Stock, Pierre ;
Joulin, Armand ;
Jegou, Herve ;
Douze, Matthijs .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :12239-12249
[17]   A Survey on Vision Transformer [J].
Han, Kai ;
Wang, Yunhe ;
Chen, Hanting ;
Chen, Xinghao ;
Guo, Jianyuan ;
Liu, Zhenhua ;
Tang, Yehui ;
Xiao, An ;
Xu, Chunjing ;
Xu, Yixing ;
Yang, Zhaohui ;
Zhang, Yiman ;
Tao, Dacheng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (01) :87-110
[18]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[19]   Densely Connected Convolutional Networks [J].
Huang, Gao ;
Liu, Zhuang ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2261-2269
[20]   Facial expression recognition with grid-wise attention and visual transformer [J].
Huang, Qionghao ;
Huang, Changqin ;
Wang, Xizhe ;
Jiang, Fan .
INFORMATION SCIENCES, 2021, 580 :35-54