SEDA: Self-ensembling ViT with Defensive Distillation and Adversarial Training for Robust Chest X-Rays Classification

被引:0
作者
Imam, Raza [1 ]
Almakky, Ibrahim [1 ]
Alrashdi, Salma [1 ]
Alrashdi, Baketah [1 ]
Yaqub, Mohammad [1 ]
机构
[1] Mohamed Bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
来源
DOMAIN ADAPTATION AND REPRESENTATION TRANSFER, DART 2023 | 2024年 / 14293卷
关键词
Ensembling; Adversarial Attack; Defensive Distillation; Adversarial Training; Vision Transformer;
D O I
10.1007/978-3-031-45857-6_13
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Learning methods have recently seen increased adoption in medical imaging applications. However, elevated vulnerabilities have been explored in recent Deep Learning solutions, which can hinder future adoption. Particularly, the vulnerability of Vision Transformer (ViT) to adversarial, privacy, and confidentiality attacks raise serious concerns about their reliability in medical settings. This work aims to enhance the robustness of self-ensembling ViTs for the tuberculosis chest x-ray classification task. We propose Self-Ensembling ViT with defensive Distillation and Adversarial training (SEDA). SEDA utilizes efficient CNN blocks to learn spatial features with various levels of abstraction from feature representations extracted from intermediate ViT blocks, that are largely unaffected by adversarial perturbations. Furthermore, SEDA leverages adversarial training in combination with defensive distillation for improved robustness against adversaries. Training using adversarial examples leads to better model generalizability and improves its ability to handle perturbations. Distillation using soft probabilities introduces uncertainty and variation into the output probabilities, making it more difficult for adversarial and privacy attacks. Extensive experiments performed with the proposed architecture and training paradigm on publicly available Tuberculosis x-ray dataset shows SOTA efficacy of SEDA compared to SEViT in terms of computational efficiency with 70x times lighter framework and enhanced robustness of +9%. Code: Github.
引用
收藏
页码:126 / 135
页数:10
相关论文
共 18 条
[1]   Self-Ensembling Vision Transformer (SEViT) for Robust Medical Image Classification [J].
Almalik, Faris ;
Yaqub, Mohammad ;
Nandakumar, Karthik .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT III, 2022, 13433 :376-386
[2]   Deep learning for chest X-ray analysis: A survey [J].
calli, Erdi ;
Sogancioglu, Ecem ;
van Ginneken, Bram ;
van Leeuwen, Kicky G. ;
Murphy, Keelin .
MEDICAL IMAGE ANALYSIS, 2021, 72 (72)
[3]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[4]  
Carlini Nicholas, 2023, arXiv, DOI [DOI 10.48550/ARXIV.2301.13188, 10.48550/arXiv.2301.13188]
[5]  
Croce F, 2020, PR MACH LEARN RES, V119
[6]  
Goodfellow I., 2015, INT C LEARNING REPRE, P1
[7]   Privacy-Preserving Deep Learning With Learnable Image Encryption on Medical Images [J].
Huang, Qi-Xian ;
Yap, Wai Leong ;
Chiu, Min-Yi ;
Sun, Hung-Min .
IEEE ACCESS, 2022, 10 :66345-66355
[8]   Adversarial attacks and defenses on AI in medical imaging informatics: A survey [J].
Kaviani, Sara ;
Han, Ki Jin ;
Sohn, Insoo .
EXPERT SYSTEMS WITH APPLICATIONS, 2022, 198
[9]  
Madry A., 2017, ARXIV
[10]  
Malik H.S., 2022, 33 BRIT MACH VIS C 2