A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability?

被引:230
|
作者
Huang, Xiaowei [1 ]
Kroening, Daniel [2 ]
Ruan, Wenjie [3 ]
Sharp, James [4 ]
Sun, Youcheng [5 ]
Thamo, Emese [1 ]
Wu, Min [2 ]
Yi, Xinping [1 ]
机构
[1] Univ Liverpool, Liverpool, Merseyside, England
[2] Univ Oxford, Oxford, England
[3] Univ Lancaster, Lancaster, England
[4] Def Sci & Technol Lab Dstl, Porton Down Salisbury, England
[5] Queens Univ Belfast, Belfast, Antrim, North Ireland
基金
英国工程与自然科学研究理事会;
关键词
ABSTRACTION-REFINEMENT; ROBUSTNESS; EXTRACTION;
D O I
10.1016/j.cosrev.2020.100270
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the past few years, significant progress has been made on deep neural networks (DNNs) in achieving human-level performance on several long-standing tasks. With the broader deployment of DNNs on various applications, the concerns over their safety and trustworthiness have been raised in public, especially after the widely reported fatal incidents involving self-driving cars. Research to address these concerns is particularly active, with a significant number of papers released in the past few years. This survey paper conducts a review of the current research effort into making DNNs safe and trustworthy, by focusing on four aspects: verification, testing, adversarial attack and defence, and interpretability. In total, we survey 202 papers, most of which were published after 2017. (c) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页数:35
相关论文
共 50 条
  • [1] Interpretability Analysis of Deep Neural Networks With Adversarial Examples
    Dong Y.-P.
    Su H.
    Zhu J.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (01): : 75 - 86
  • [2] ADVERSARIAL WATERMARKING TO ATTACK DEEP NEURAL NETWORKS
    Wang, Gengxing
    Chen, Xinyuan
    Xu, Chang
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1962 - 1966
  • [3] Safety Verification of Deep Neural Networks
    Huang, Xiaowei
    Kwiatkowska, Marta
    Wang, Sen
    Wu, Min
    COMPUTER AIDED VERIFICATION, CAV 2017, PT I, 2017, 10426 : 3 - 29
  • [4] Cocktail Universal Adversarial Attack on Deep Neural Networks
    Li, Shaoxin
    Li, Xiaofeng
    Che, Xin
    Li, Xintong
    Zhang, Yong
    Chu, Lingyang
    COMPUTER VISION - ECCV 2024, PT LXV, 2025, 15123 : 396 - 412
  • [5] Diversity Adversarial Training against Adversarial Attack on Deep Neural Networks
    Kwon, Hyun
    Lee, Jun
    SYMMETRY-BASEL, 2021, 13 (03):
  • [6] Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing Their Input Gradients
    Ros, Andrew Slavin
    Doshi-Velez, Finale
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 1660 - 1669
  • [7] ADMM Attack: An Enhanced Adversarial Attack for Deep Neural Networks with Undetectable Distortions
    Zhao, Pu
    Xu, Kaidi
    Liu, Sijia
    Wang, Yanzhi
    Lin, Xue
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 499 - 505
  • [8] AdvAttackVis: An Adversarial Attack Visualization System for Deep Neural Networks
    Ding Wei-jie
    Shen Xuchen
    Yuan Ying
    Mao Ting-yun
    Sun Guo-dao
    Chen Li-li
    Chen Bing-ting
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (05) : 383 - 391
  • [9] Adversarial attack model based on deep neural network interpretability and artificial fish swarm algorithm
    Li, Yamin
    INTERNATIONAL JOURNAL OF ELECTRONIC SECURITY AND DIGITAL FORENSICS, 2024, 16 (05) : 614 - 632
  • [10] Survey on Testing of Deep Neural Networks
    Wang Z.
    Yan M.
    Liu S.
    Chen J.-J.
    Zhang D.-D.
    Wu Z.
    Chen X.
    Ruan Jian Xue Bao/Journal of Software, 2020, 31 (05): : 1255 - 1275