Explainable Deep Learning to Classify Royal Navy Ships

被引:2
作者
Baesens, Bart [1 ,2 ]
Adams, Amy [3 ]
Pacheco-Ruiz, Rodrigo [3 ]
Baesens, Ann-Sophie [1 ]
Broucke, Seppe Vanden [4 ]
机构
[1] Katholieke Univ Leuven, Res Ctr Informat Syst Engn LIRIS, B-3000 Leuven, Belgium
[2] Univ Southampton, Sch Management, Southampton SO17 1BJ, Hants, England
[3] Natl Museum Royal Navy, Portsmouth PO1 3NH, Hants, England
[4] UGent, Dept Business Informat & Operat Management, B-9000 Ghent, Belgium
关键词
Convolutional neural networks; deep learning; explainability; digitised archives; image classification; royal navy; CLASSIFICATION; IMAGES;
D O I
10.1109/ACCESS.2023.3346061
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We research how deep learning convolutional neural networks can be used to automatically classify the unique data set of black-and-white naval ships images from the Wright and Logan photographic collection held by the National Museum of the Royal Navy. We contrast various types of deep learning methods: pretrained models such as ConvNeXt, ResNet and EfficientNet, and ConvMixer. We also thoroughly investigate the impact of data preprocessing and externally obtained images on model performance. Finally, we research how the models estimated can be made transparent using visually appealing interpretability techniques such as Grad-CAM. We find that ConvNeXt has the best performance for our data set achieving an accuracy of 79.62% for 0-notch classification and an impressive 94.86% for 1-notch classification. The results indicate the importance of appropriate image preprocessing. Image segmentation combined with soft augmentation significantly contributes to model performance. We consider this research to be original in several aspects. Notably, it distinguishes itself through the uniqueness of the acquired dataset. Additionally, its distinctiveness extends to the analytical modeling pipeline, which encompasses a comprehensive range of modeling steps, including data preprocessing (incorporating external data, image segmentation, and image augmentation) and the use of deep learning techniques such as ConvNeXt, ResNet, EfficientNet, and ConvMixer. Furthermore, the research employs explanatory tools like Grad-CAM to enhance model interpretability and usability. We believe the proposed methodology offers lots of potential for documenting historic image collections.
引用
收藏
页码:1774 / 1785
页数:12
相关论文
共 24 条
[1]   Integrated use of KOS and deep learning for data set annotation in tourism domain [J].
Aracri, Giovanna ;
Folino, Antonietta ;
Silvestri, Stefano .
JOURNAL OF DOCUMENTATION, 2023, 79 (06) :1440-1458
[2]  
Cushing A. L., 2023, J. Document, V79, P1229
[3]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[4]  
Diaz-Rodriguez Natalia, 2020, UMAP '20 Adjunct: Adjunct Publication of the 28th ACM Conference on User Modeling, Adaptation and Personalization, P317, DOI 10.1145/3386392.3399276
[5]  
Goodfellow I, 2016, ADAPT COMPUT MACH LE, P1
[6]   Classification of inbound and outbound ships using convolutional neural networks [J].
Guo, Doudou ;
Gao, Dazhi ;
Chen, Zhuo ;
Li, Yuzheng ;
Zhao, Xiaojing ;
Song, Wenhua ;
Li, Xiaolei .
FRONTIERS IN MARINE SCIENCE, 2023, 10
[7]   Ship Classification from SAR Images Based on Deep Learning [J].
Hashimoto, Shintaro ;
Sugimoto, Yohei ;
Hamamoto, Ko ;
Ishihama, Naoki .
INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 1, 2019, 868 :18-34
[8]  
Hazarika Abhinaba, 2022, Computer Vision and Image Processing: 6th International Conference, CVIP 2021, Revised Selected Papers. Communications in Computer and Information Science (1568), P60, DOI 10.1007/978-3-031-11349-9_6
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]   Densely Connected Convolutional Networks [J].
Huang, Gao ;
Liu, Zhuang ;
van der Maaten, Laurens ;
Weinberger, Kilian Q. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2261-2269