The Role of 'Sign' and 'Direction' of Gradient on the Performance of CNN

被引:6
作者
Agarwal, Akshay [1 ]
Singh, Richa [2 ]
Vatsa, Mayank [2 ]
机构
[1] IIIT Delhi, Delhi, India
[2] IIT Jodhpur, Jodhpur, Rajasthan, India
来源
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020) | 2020年
关键词
D O I
10.1109/CVPRW50498.2020.00331
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
State-of-the-art deep learning models have achieved superlative performance across multiple computer vision applications such as object recognition, face recognition, and digits/character classification. Most of these models highly rely on the gradient information flows through the network for learning. By utilizing this gradient information, a simple gradient sign method based attack is developed to fool the deep learning models. However, the primary concern with this attack is the perceptibility of noise for large degradation in classification accuracy. This research address the question of whether an imperceptible gradient noise can be generated to fool the deep neural networks? For this, the role of sign function in the gradient attack is analyzed. The analysis shows that without-sign function, i.e. gradient magnitude, not only leads to a successful attack mechanism but the noise is also imperceptible to the human observer. Extensive quantitative experiments performed using two convolutional neural networks validate the above observation. For instance, AlexNet architecture yields 63.54% accuracy on the CIFAR-10 database which reduces to 0.0% and 26.39% when sign (i.e., perceptible) and without-sign (i.e., imperceptible) of the gradient is utilized, respectively. Further, the role of the direction of the gradient for image manipulation is studied. When an image is manipulated in the positive direction of the gradient, an adversarial image is generated. On the other hand, if the opposite direction of the gradient is utilized for image manipulation, it is observed that the classification error rate of the CNN model is reduced. On AlexNet, the error rate of 36.46% reduces to 4.29% when images of CIFAR-10 are manipulated in the negative direction of the gradient. To explore other enthusiastic results on multiple object databases, including CIFAR-100, fashion-MNIST, and SVHN, please refer to the full paper.
引用
收藏
页码:2748 / 2756
页数:9
相关论文
共 29 条
  • [1] Improving Spatial Resolution Using Incoherent Subtraction of Receive Beams Having Different Apodizations
    Agarwal, Anil
    Reeg, Jonathan
    Podkowa, Anthony S.
    Oelze, Michael L.
    [J]. IEEE TRANSACTIONS ON ULTRASONICS FERROELECTRICS AND FREQUENCY CONTROL, 2019, 66 (01) : 5 - 17
  • [2] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [3] Chen PY, 2018, AAAI CONF ARTIF INTE, P10
  • [4] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [5] Texture and art with deep neural networks
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2017, 46 : 178 - 186
  • [6] DeepRing: Protecting Deep Neural Network with Blockchain
    Goel, Akhil
    Agarwal, Akshay
    Vatsa, Mayank
    Singh, Richa
    Ratha, Nalini
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 2821 - 2828
  • [7] Goel S, 2019, FAMILY FIRMS AND INSTITUTIONAL CONTEXTS: BUSINESS MODELS, INNOVATION AND COMPETITIVE ADVANTAGE, P1
  • [8] Goodfellow I.J., 2015, INT C LEARN REPR ICL, P1
  • [9] Detecting and Mitigating Adversarial Perturbations for Robust Face Recognition
    Goswami, Gaurav
    Agarwal, Akshay
    Ratha, Nalini
    Sing, Richa
    Vatsa, Mayank
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2019, 127 (6-7) : 719 - 742
  • [10] Goswami G, 2018, AAAI CONF ARTIF INTE, P6829