Deep learning interpretability analysis methods in image interpretation

被引:0
作者
Gong J. [1 ,2 ]
Huan L. [1 ]
Zheng X. [1 ]
机构
[1] State Key Laboratory of Information Engineering in Surveying, Mapping and Remoto Sensing, Wuhan University, Wuhan
[2] School of Remote Sensing and Engineering, Wuhan University, Wuhan
来源
Cehui Xuebao/Acta Geodaetica et Cartographica Sinica | 2022年 / 51卷 / 06期
基金
中国国家自然科学基金;
关键词
artificial intelligence; deep learning; interpretability; remote sensing interpretation; review;
D O I
10.11947/j.AGCS.2022.20220106
中图分类号
学科分类号
摘要
The rapid development of deep learning has greatly improved the performance of various computer vision tasks. However, the "black box" nature of deep learning network models makes it difficult for users to understand its decision-making mechanism, which is not conductive to model structure optimization and security enhancement and also greatly increases the training cost. Focusing on the task of intelligent image interpretation, this paper makes a comprehensive review and comparison of the research progress of deep learning interpretability. Firstly, we group the current interpretability analysis methods into six categories: activation maximization method, surrogate model, attribution method, perturbation-based method, class activation map based method and example-based method, and review the principle, focus, advantages, and disadvantages of existing related works. Secondly, we introduce eight evaluation metrics that measure the reliability of the explanations provided by the various interpretability analysis methods, and sort out the current publicly available open source libraries for deep learning interpretability analysis. Based on the open source library, we verify the applicability of the current deep learning interpretability analysis methods to the interpretation of remote sensing images. The experimental results show that the current interpretability methods are applicable to the analysis of remote sensing interpretation, but have certain limitations. Finally, we summarize the open challenges of using existing interpretability algorithms for remote sensing data analysis, and look forward to the prospect of designing interpretability analysis methods oriented to remote sensing images. We hope this review can promote the research on interpretability methods for remote sensing image interpretation, so as to provide reliable theoretical support and algorithm design guidance for the application of deep learning technology in remote sensing image interpretation tasks. © 2022 SinoMaps Press. All rights reserved.
引用
收藏
页码:873 / 884
页数:11
相关论文
共 68 条
  • [1] LAPUSCHKIN S, WALDCHEN S, BINDER A, Et al., Unmasking clever hans predictors and assessing what machines really learn[J], Nature Communications, 10, 1, (2019)
  • [2] EVERINGHAM M, VAN GOOL L, WILLIAMS C K I, Et al., The PASCAL Visual Object Classes (VOC) challenge[J], International Journal of Computer Vision, 88, 2, pp. 303-338, (2010)
  • [3] RIBEIRO M T, SINGH S, GUESTRIN C., Why should I trust you?":explaining the predictions of any classifier, Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 1135-1144, (2016)
  • [4] GEIRHOS R, RUBISCH P, MICHAELIS C, Et al., ImageNet-trained CNNs are biased towards texture
  • [5] increasing shape bias improves accuracy and robustness, Proceedings of the 7th International Conference on Learning Representations, (2019)
  • [6] NGUYEN A, YOSINSKI J, CLUNE J., Understanding neural networks via feature visualization:a survey, Explainable AI:Interpreting, Explaining and Visualizing Deep Learning, pp. 55-76, (2019)
  • [7] ERHAN D, BENGIO Y, COURVILLE A, Et al., Visualizing higher-layer features of a deep network, (2009)
  • [8] NGUYEN A, YOSINSKI J, CLUNE J., Deep neural networks are easily fooled:high confidence predictions for unrecognizable images, Proceedings of 2015 IEEE Conference on Computer Vision and Pattern Recognition, pp. 427-436, (2015)
  • [9] SIMONYAN K, VEDALDI A, ZISSERMAN A., Deep inside convolutional networks:Visualising image classification models and saliency maps, Proceedings of the 2nd International Conference on Learning Representations, (2014)
  • [10] MAHENDRAN A, VEDALDI A., Visualizing deep convolutional neural networks using natural pre-images[J], International Journal of Computer Vision, 120, 3, pp. 233-255, (2016)