LAFED: Towards robust ensemble models via Latent Feature Diversification

被引:2
作者
Zhuang, Wenzi [1 ]
Huang, Lifeng [2 ]
Gao, Chengying [1 ]
Liu, Ning [1 ,3 ]
机构
[1] Sun Yet Sen Univ, Sch Comp Sci & Engn, Guangzhou, Guangdong, Peoples R China
[2] South China Agr Univ, Coll Math & Informat, Guangzhou, Guangdong, Peoples R China
[3] Guangdong Key Lab Informat Secur Technol, Guangzhou, Peoples R China
关键词
Adversarial example; Adversarial defense; Ensemble model; Robustness;
D O I
10.1016/j.patcog.2023.110225
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial examples pose a significant challenge to the security of deep neural networks (DNNs). In order to defend against malicious attacks, adversarial training forces DNNs to learn more robust features by suppressing generalizable but non -robust features, which boosts the robustness while suffering from significant accuracy drops on clean images. Ensemble training, on the other hand, trains multiple sub -models to predict data for improved robustness and still achieves desirable accuracy on clean data. Despite these efforts, previous ensemble methods are still susceptible to attacks and fail to increase model diversity as the size of the ensemble group increases. In this work, we revisit the model diversity from the perspective of data and discover that high similarity between training batches decreases feature diversity and weakens ensemble robustness. To this end, we propose Latent Feature Diversification (LAFED), which reconstructs training sets with diverse features during the optimization, enhancing the overall robustness of an ensemble. For each sub -model, LAFED treats the vulnerability extracted from other sub -models as raw data, which is then combined with round -changed weights with a stochastic manner in the latent space. This results in the formation of new features, remarkably reducing the similarity of learned representations between the submodels. Furthermore, LAFED enhances feature diversity within the ensemble model by utilizing hierarchical smoothed labels. Extensive experiments illustrate that LAFED significantly improves diversity among submodels and enhances robustness against adversarial attacks compared to current methods. The code is publicly available at https://github.com/zhuangwz/LAFED.
引用
收藏
页数:13
相关论文
共 42 条
[1]  
Atzmon M, 2019, ADV NEUR IN, V32
[2]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[3]   Towards Evaluating the Robustness of Neural Networks [J].
Carlini, Nicholas ;
Wagner, David .
2017 IEEE SYMPOSIUM ON SECURITY AND PRIVACY (SP), 2017, :39-57
[4]   A black-Box adversarial attack for poisoning clustering [J].
Cina, Antonio Emanuele ;
Torcinovich, Alessandro ;
Pelillo, Marcello .
PATTERN RECOGNITION, 2022, 122
[5]  
Croce F, 2020, PR MACH LEARN RES, V119
[6]   Ensemble methods in machine learning [J].
Dietterich, TG .
MULTIPLE CLASSIFIER SYSTEMS, 2000, 1857 :1-15
[7]   Adversarial Laser Beam: Effective Physical-World Attack to DNNs in a Blink [J].
Duan, Ranjie ;
Mao, Xiaofeng ;
Qin, A. K. ;
Chen, Yuefeng ;
Ye, Shaokai ;
He, Yuan ;
Yang, Yun .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :16057-16066
[8]   Adversarial Camouflage: Hiding Physical-World Attacks with Natural Styles [J].
Duan, Ranjie ;
Ma, Xingjun ;
Wang, Yisen ;
Bailey, James ;
Qin, A. K. ;
Yang, Yun .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :997-1005
[9]  
Fu CH, 2020, Arxiv, DOI [arXiv:2009.08233, 10.48550/arXiv.2009.08233]
[10]   A black-box adversarial attack strategy with adjustable sparsity and generalizability for deep image classifiers [J].
Ghosh, Arka ;
Mullick, Sankha Subhra ;
Datta, Shounak ;
Das, Swagatam ;
Das, Asit Kr ;
Mallipeddi, Rammohan .
PATTERN RECOGNITION, 2022, 122