Image Transformation-Based Defense Against Adversarial Perturbation on Deep Learning Models

被引:34
作者
Agarwal, Akshay [1 ]
Singh, Richa [2 ]
Vatsa, Mayank [2 ]
Ratha, Nalini K. [3 ]
机构
[1] IIIT Delhi, Dept Comp Sci & Engn, Delhi 110020, India
[2] IIT Jodhpur, Dept Comp Sci & Engn, Karwar 342037, Rajasthan, India
[3] SUNY Buffalo, Buffalo, NY 14260 USA
关键词
Perturbation methods; Databases; Machine learning; Transforms; Detection algorithms; Integrated circuits; Machine learning algorithms; Adversarial attack; detection; mitigation; image transformations; deep learning; NEURAL-NETWORKS; ROBUST;
D O I
10.1109/TDSC.2020.3027183
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning algorithms provide state-of-the-art results on a multitude of applications. However, it is also well established that they are highly vulnerable to adversarial perturbations. It is often believed that the solution to this vulnerability of deep learning systems must come from deep networks only. Contrary to this common understanding, in this article, we propose a non-deep learning approach that searches over a set of well-known image transforms such as Discrete Wavelet Transform and Discrete Sine Transform, and classifying the features with a support vector machine-based classifier. Existing deep networks-based defense have been proven ineffective against sophisticated adversaries, whereas image transformation-based solution makes a strong defense because of the non-differential nature, multiscale, and orientation filtering. The proposed approach, which combines the outputs of two transforms, efficiently generalizes across databases as well as different unseen attacks and combinations of both (i.e., cross-database and unseen noise generation CNN model). The proposed algorithm is evaluated on large scale databases, including object database (validation set of ImageNet) and face recognition (MBGC) database. The proposed detection algorithm yields at-least 84.2% and 80.1% detection accuracy under seen and unseen database test settings, respectively. Besides, we also show how the impact of the adversarial perturbation can be neutralized using a wavelet decomposition-based filtering method of denoising. The mitigation results with different perturbation methods on several image databases demonstrate the effectiveness of the proposed method.
引用
收藏
页码:2106 / 2121
页数:16
相关论文
共 106 条
  • [1] Ackerman E., 2017, IEEE Spectr, V1
  • [2] Noise is Inside Me! Generating Adversarial Perturbations with Noise Derived from Natural Filters
    Agarwal, Akshay
    Vatsa, Mayank
    Singh, Richa
    Ratha, Nalini K.
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 3354 - 3363
  • [3] Agarwal A., 2018, P IEEE 9 INT C BIOM, P1
  • [4] The Role of 'Sign' and 'Direction' of Gradient on the Performance of CNN
    Agarwal, Akshay
    Singh, Richa
    Vatsa, Mayank
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 2748 - 2756
  • [5] Defense against Universal Adversarial Perturbations
    Akhtar, Naveed
    Liu, Jian
    Mian, Ajmal
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 3389 - 3398
  • [6] Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey
    Akhtar, Naveed
    Mian, Ajmal
    [J]. IEEE ACCESS, 2018, 6 : 14410 - 14430
  • [7] [Anonymous], 2017, AISEC
  • [8] [Anonymous], 2018, P ICLR
  • [9] [Anonymous], 2017, P ACM WORKSH ART INT, DOI DOI 10.1145/3128572.3140449
  • [10] [Anonymous], 2016, ARXIV161101236