2N labeling defense method against adversarial attacks by filtering and extended class label set

被引:1
作者
Szucs, Gabor [1 ]
Kiss, Richard [1 ]
机构
[1] Budapest Univ Technol & Econ, Dept Telecommun & Media Informat, Magyar Tudosok Krt 2, H-1117 Budapest, Hungary
关键词
Adversarial attack; Deep learning; Defense method; Filtering; Image classification;
D O I
10.1007/s11042-022-14021-5
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The fast improvement of deep learning methods resulted in breakthroughs in image classification, however, these models are sensitive to adversarial perturbations, which can cause serious problems. Adversarial attacks try to change the model output by adding noise to the input, in our research we propose a combined defense method against it. Two defense approaches have been evolved in the literature, one robustizes the attacked model for higher accuracy, and the other approach detects the adversarial examples. Only very few papers discuss both approaches, thus our aim was to combine them to obtain a more robust model and to examine the combination, in particular the filtering capability of the detector. Our contribution was that the filtering based on the decision of the detector is able to enhance the accuracy, which was theoretically proved. Besides that, we developed a novel defense method called 2N labeling, where we extended the idea of the NULL labeling method. While the NULL labeling suggests only one new class for the adversarial examples, the 2N labeling method suggests twice as much. The novelty of our idea is that a new extended class is assigned to each original class, as the adversarial version of it, thus it assists the detector and robust classifier as well. The 2N labeling method was compared to competitor methods on two test datasets. The results presented that our method surpassed the others, and it can operate with a constant classification performance regardless of the presence or amplitude of adversarial attacks.
引用
收藏
页码:16717 / 16740
页数:24
相关论文
共 54 条
  • [1] Abdu-Aguye MG, 2020, INT CONF ACOUST SPEE, P3092, DOI [10.1109/ICASSP40776.2020.9053311, 10.1109/icassp40776.2020.9053311]
  • [2] An adversarial attack detection method in deep neural networks based on re-attacking approach
    Ahmadi, Morteza Ali
    Dianat, Rouhollah
    Amirkhani, Hossein
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (07) : 10985 - 11014
  • [3] Alparslan Y, 2020, ARXIV
  • [4] Brendel W., 2017, arXiv
  • [5] Breve B, 2020, ITASEC, P71
  • [6] Towards Evaluating the Robustness of Neural Networks
    Carlini, Nicholas
    Wagner, David
    [J]. 2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, : 39 - 57
  • [7] GDPR Compliant Information Confidentiality Preservation in Big Data Processing
    Caruccio, Loredana
    Desiato, Domenico
    Polese, Giuseppe
    Tortora, Genoveffa
    [J]. IEEE ACCESS, 2020, 8 (08): : 205034 - 205050
  • [8] Social network data analysis to highlight privacy threats in sharing data
    Cerruto, Francesca
    Cirillo, Stefano
    Desiato, Domenico
    Gambardella, Simone Michele
    Polese, Giuseppe
    [J]. JOURNAL OF BIG DATA, 2022, 9 (01)
  • [9] A survey on adversarial attacks and defences
    Chakraborty, Anirban
    Alam, Manaar
    Dey, Vishal
    Chattopadhyay, Anupam
    Mukhopadhyay, Debdeep
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2021, 6 (01) : 25 - 45
  • [10] HopSkipJumpAttack: A Query-Efficient Decision-Based Attack
    Chen, Jianbo
    Jordan, Michael, I
    Wainwright, Martin J.
    [J]. 2020 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP 2020), 2020, : 1277 - 1294