Adversarial Defense on Harmony: Reverse Attack for Robust AI Models Against Adversarial Attacks

被引:0
|
作者
Kim, Yebon [1 ]
Jung, Jinhyo [2 ]
Kim, Hyunjun [1 ]
So, Hwisoo [2 ,3 ]
Ko, Yohan [4 ]
Shrivastava, Aviral [3 ]
Lee, Kyoungwoo [2 ]
Hwang, Uiwon [5 ]
机构
[1] Yonsei Univ, Dept Comp Sci, Wonju 26493, South Korea
[2] Yonsei Univ, Dept Comp Sci, Seoul 03722, South Korea
[3] Arizona State Univ, Sch Comp & Augmented Intelligence, Tempe, AZ 85281 USA
[4] Yonsei Univ, Div Software, Wonju 26493, South Korea
[5] Yonsei Univ, Div Digital Healthcare, Wonju 26493, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
基金
新加坡国家研究基金会; 美国国家科学基金会;
关键词
Training; Perturbation methods; Accuracy; Noise; Mathematical models; Robustness; Computational modeling; Closed box; Glass box; Deep learning; Deep neural networks; adversarial attacks and defenses; security; reliability;
D O I
10.1109/ACCESS.2024.3505215
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) are crucial in safety-critical applications but vulnerable to adversarial attacks, where subtle perturbations cause misclassification. Existing defense mechanisms struggle with small perturbations and face accuracy-robustness trade-offs. This study introduces the "Reverse Attack" method to address these challenges. Our approach uniquely reconstructs and classifies images by applying perturbations opposite to the attack direction, using a complementary "Revenant" classifier to maintain original image accuracy. The proposed method significantly outperforms existing strategies, maintaining clean image accuracy with only a 2.92% decrease while achieving over 70% robust accuracy against all benchmarked adversarial attacks. This contrasts with current mechanisms, which typically suffer an 18% reduction in clean image accuracy and only 36% robustness against adversarial examples. We evaluate our method on the CIFAR-10 dataset using ResNet50, testing against various attacks including PGD and components of Auto Attack. Although our approach incurs additional computational costs during reconstruction, our method represents a significant advancement in robust defenses against adversarial attacks while preserving clean input performance. This balanced approach paves the way for more reliable DNNs in critical applications. Future work will focus on optimization and exploring applicability to larger datasets and complex architectures.
引用
收藏
页码:176485 / 176497
页数:13
相关论文
共 50 条
  • [1] Text Adversarial Purification as Defense against Adversarial Attacks
    Li, Linyang
    Song, Demin
    Qiu, Xipeng
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 338 - 350
  • [2] Towards Robust Ensemble Defense Against Adversarial Examples Attack
    Mani, Nag
    Moh, Melody
    Moh, Teng-Sheng
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [3] Attack-less adversarial training for a robust adversarial defense
    Ho, Jiacang
    Lee, Byung-Gook
    Kang, Dae-Ki
    APPLIED INTELLIGENCE, 2022, 52 (04) : 4364 - 4381
  • [4] Attack-less adversarial training for a robust adversarial defense
    Jiacang Ho
    Byung-Gook Lee
    Dae-Ki Kang
    Applied Intelligence, 2022, 52 : 4364 - 4381
  • [5] Deblurring as a Defense against Adversarial Attacks
    Duckworth, William, III
    Liao, Weixian
    Yu, Wei
    2023 IEEE 12TH INTERNATIONAL CONFERENCE ON CLOUD NETWORKING, CLOUDNET, 2023, : 61 - 67
  • [6] Blind Adversarial Training: Towards Comprehensively Robust Models Against Blind Adversarial Attacks
    Xie, Haidong
    Xiang, Xueshuang
    Dong, Bin
    Liu, Naijin
    ARTIFICIAL INTELLIGENCE, CICAI 2023, PT II, 2024, 14474 : 15 - 26
  • [7] ROLDEF: RObust Layered DEFense for Intrusion Detection Against Adversarial Attacks
    Gungor, Onat
    Rosing, Tajana
    Alcsanli, Bans
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [8] Are Malware Detection Models Adversarial Robust Against Evasion Attack?
    Rathore, Hemant
    Samavedhi, Adithya
    Sahay, Sanjay K.
    Sewak, Mohit
    IEEE INFOCOM 2022 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (INFOCOM WKSHPS), 2022,
  • [9] The Best Defense is a Good Offense: Adversarial Augmentation against Adversarial Attacks
    Frosio, Iuri
    Kautz, Jan
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 4067 - 4076
  • [10] Defense Against Adversarial Attacks Using Topology Aligning Adversarial Training
    Kuang, Huafeng
    Liu, Hong
    Lin, Xianming
    Ji, Rongrong
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 3659 - 3673