Clean-label attack based on negative afterimage on neural networks

被引:0
作者
Zang, Liguang [1 ]
Li, Yuancheng [1 ]
机构
[1] North China Elect Power Univ, Dept Control & Comp Engineeringg, Beijing 102206, Peoples R China
关键词
Clean-label attack; Adversarial attacks; Afterimage; Machine learning;
D O I
10.1007/s13042-024-02230-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial attacks can fool machine learning models by adding small but carefully designed perturbations to an image, resulting in the model to misclassify it into another class. In general, an attacker chooses a target class as a pollution source to generate a perturbed image and tricks the classifier into classifying it as the target class. However, the effectiveness of adversarial attacks suffers from some limitations, including the need for specific control over the labeling of the target class and challenges in generating effective adversarial samples in real-world scenarios. To overcome these limitations, we propose a novel clean-label attack based on negative afterimage. Negative afterimage is a visual phenomenon whereby, after gazing at a bright image for some time, one perceives the inverse color image of the previously observed bright region when shifting focus to a darker area. We explore the afterimage phenomenon and use negative afterimages as pollution sources to generate adversarial samples, which can avoid any dependency on the labeling of pollution sources. Subsequently, we generate adversarial samples using an optimization-based approach to ensure that adversarial samples are visually undetectable. We conducted experiments on two widely used datasets, CIFAR-10 and ImageNet. The results show that our proposed method can effectively generate adversarial samples with high stealthiness.
引用
收藏
页码:449 / 460
页数:12
相关论文
共 50 条
[41]   Multi-Targeted Backdoor: Indentifying Backdoor Attack for Multiple Deep Neural Networks [J].
Kwon, Hyun ;
Yoon, Hyunsoo ;
Park, Ki-Woong .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2020, E103D (04) :883-887
[42]   Power Analysis Attack Using Neural Networks with Wavelet Transform as Pre-processor [J].
Saravanan, P. ;
Kalpana, P. ;
Preethisri, V. ;
Sneha, V. .
18TH INTERNATIONAL SYMPOSIUM ON VLSI DESIGN AND TEST, 2014,
[43]   Detection of Unknown DDoS Attack Using Convolutional Neural Networks Featuring Geometrical Metric [J].
Shieh, Chin-Shiuh ;
Nguyen, Thanh-Tuan ;
Horng, Mong-Fong .
MATHEMATICS, 2023, 11 (09)
[44]   Sybil Attack Detection Based on Signal Clustering in Vehicular Networks [J].
Tulay, Halit Bugra ;
Koksal, Can Emre .
IEEE TRANSACTIONS ON MACHINE LEARNING IN COMMUNICATIONS AND NETWORKING, 2024, 2 :753-765
[45]   BoW-based neural networks vs. cutting-edge models for single-label text classification [J].
Hassan I. Abdalla ;
Ali A. Amer ;
Sri Devi Ravana .
Neural Computing and Applications, 2023, 35 :20103-20116
[46]   BoW-based neural networks vs. cutting-edge models for single-label text classification [J].
Abdalla, Hassan I. ;
Amer, Ali A. ;
Ravana, Sri Devi .
NEURAL COMPUTING & APPLICATIONS, 2023, 35 (27) :20103-20116
[47]   MIMLRBF: RBF neural networks for multi-instance multi-label learning [J].
Zhang, Min-Ling ;
Wang, Zhi-Jian .
NEUROCOMPUTING, 2009, 72 (16-18) :3951-3956
[48]   Multi-label classification of line chart images using convolutional neural networks [J].
Kosemen, Cem ;
Birant, Derya .
SN APPLIED SCIENCES, 2020, 2 (07)
[49]   Multi-label classification of line chart images using convolutional neural networks [J].
Cem Kosemen ;
Derya Birant .
SN Applied Sciences, 2020, 2
[50]   DI-AA: An interpretable white-box attack for fooling deep neural networks [J].
Wang, Yixiang ;
Liu, Jiqiang ;
Chang, Xiaolin ;
Rodriguez, Ricardo J. ;
Wang, Jianhua .
INFORMATION SCIENCES, 2022, 610 :14-32