Similarity-driven adversarial testing of neural networks

被引:0
|
作者
Filus, Katarzyna [1 ]
Domanska, Joanna [1 ]
机构
[1] Polish Acad Sci, Inst Theoret & Appl Informat, Gliwice, Poland
关键词
Adversarial attacks; Testing; Artificial intelligence security; Convolutional Neural Networks; Object recognition;
D O I
10.1016/j.knosys.2024.112621
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although Convolutional Neural Networks (CNNs) are among the most important algorithms of computer vision and the artificial intelligence-based systems, they are vulnerable to adversarial attacks. Such attacks can cause dangerous consequences in real-life deployments. Consequently, testing of the artificial intelligence-based systems from their perspective is crucial to reliably support human prediction and decision-making through computation techniques under varying conditions. While proposing new effective attacks is important for neural network testing, it is also crucial to design effective strategies that can be used to choose target labels for these attacks. That is why, in this paper we propose a novel similarity-driven adversarial testing methodology for target label choosing. Our motivation is that CNNs, similarly to humans, tend to make mistakes mostly among categories they perceive similar. Thus, the effort to make models predict a particular class is not equal for all classes. Motivated by this, we propose to use the most and least similar labels to the ground truth according to different similarity measures to choose the target label for an adversarial attack. They can be treated as best- and worst-case scenarios in practical and transparent testing methodologies. As similarity is one of the key components of human cognition and categorization, the approach presents a shift towards amore human- centered security testing of deep neural networks. The obtained numerical results show the superiority of the proposed methods to the existing strategies in the targeted and the non-targeted testing setups.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Generation of Realistic Navigation Paths for Web Site Testing Using Recurrent Neural Networks and Generative Adversarial Neural Networks
    Pavanetto, Silvio
    Brambilla, Marco
    WEB ENGINEERING, ICWE 2020, 2020, 12128 : 244 - 258
  • [32] Similarity-driven multi-view embeddings from high-dimensional biomedical data
    Avants, Brian B.
    Tustison, Nicholas J.
    Stone, James R.
    NATURE COMPUTATIONAL SCIENCE, 2021, 1 (02): : 143 - 152
  • [33] Data-driven discovery of self-similarity using neural networks
    Watanabe, Ryota
    Ishii, Takanori
    Hirono, Yuji
    Maruoka, Hirokazu
    PHYSICAL REVIEW E, 2025, 111 (02)
  • [34] A triplet graph convolutional network with attention and similarity-driven dictionary learning for remote sensing image retrieval
    Regan, Jacob
    Khodayar, Mahdi
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 232
  • [35] Increasing Adversarial Uncertainty to Scale Private Similarity Testing
    Hua, Yiqing
    Namavari, Armin
    Cheng, Kaishuo
    Naaman, Mor
    Ristenpart, Thomas
    PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 1777 - 1794
  • [36] Similarity of Neural Architectures Using Adversarial Attack Transferability
    Hwang, Jaehui
    Han, Dongyoon
    Heo, Byeongho
    Park, Song
    Chun, Sanghyuk
    Lee, Jong-Seok
    COMPUTER VISION - ECCV 2024, PT XXXVII, 2025, 15095 : 106 - 126
  • [37] Similarity learning with neural networks
    Sanfins, G.
    Ramos, F.
    Naiff, D.
    PHYSICAL REVIEW E, 2025, 111 (02)
  • [38] AVLaughterCycleEnabling a virtual agent to join in laughing with a conversational partner using a similarity-driven audiovisual laughter animation
    Jérôme Urbain
    Radoslaw Niewiadomski
    Elisabetta Bevacqua
    Thierry Dutoit
    Alexis Moinet
    Catherine Pelachaud
    Benjamin Picart
    Joëlle Tilmanne
    Johannes Wagner
    Journal on Multimodal User Interfaces, 2010, 4 : 47 - 58
  • [39] DeepPath: Path-driven Testing Criteria for Deep Neural Networks
    Wang, Dong
    Wang, Ziyuan
    Fang, Chunrong
    Chen, Yanshan
    Chen, Zhenyu
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE TESTING (AITEST), 2019, : 119 - 120
  • [40] Adversarial Dropout for Recurrent Neural Networks
    Park, Sungrae
    Song, Kyungwoo
    Ji, Mingi
    Lee, Wonsung
    Moon, Il-Chul
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 4699 - 4706