DunDi: Improving Robustness of Neural Networks Using Distance Metric Learning

被引:1
作者
Cui, Lei [1 ]
Xi, Rongrong [1 ]
Hao, Zhiyu [1 ]
Yu, Xuehao [2 ]
Zhang, Lei [1 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] State Grid Informat & Telecommun Branch, Beijing, Peoples R China
来源
COMPUTATIONAL SCIENCE - ICCS 2019, PT II | 2019年 / 11537卷
基金
北京市自然科学基金; 中国国家自然科学基金;
关键词
Robustness; Deep neural network; Metric learning;
D O I
10.1007/978-3-030-22741-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The deep neural networks (DNNs), although highly accurate, are vulnerable to adversarial attacks. A slight perturbation applied to a sample may lead to misprediction of the DNN, even it is imperceptible to humans. This defect makes the DNN lack of robustness to malicious perturbations, and thus limits their usage in many safety-critical systems. To this end, we present DunDi, a metric learning based classification model, to provide the ability to defend adversarial attacks. The key idea behind DunDi is a metric learning model which is able to pull samples of the same label together meanwhile pushing samples of different labels away. Consequently, the distance between samples and model's boundary can be enlarged accordingly, so that significant perturbations are required to fool the model. Then, based on the distance comparison, we propose a two-step classification algorithm that performs efficiently for multi-class classification. DunDi can not only build and train a new customized model but also support the incorporation of the available pre-trained neural network models to take full advantage of their capabilities. The results show that DunDi is able to defend 94.39% and 88.91% of adversarial samples generated by four state-of-the-art adversarial attacks on the MNIST dataset and CIFAR-10 dataset, without hurting classification accuracy.
引用
收藏
页码:145 / 159
页数:15
相关论文
共 50 条
[1]   Toward Certified Robustness of Distance Metric Learning [J].
Yang, Xiaochen ;
Guo, Yiwen ;
Dong, Mingzhi ;
Xue, Jing-Hao .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) :3834-3844
[2]   Unrestricted deep metric learning using neural networks interaction [J].
Mehralian, Soheil ;
Teshnehlab, Mohammad ;
Nasersharif, Babak .
PATTERN ANALYSIS AND APPLICATIONS, 2021, 24 (04) :1699-1711
[3]   Unrestricted deep metric learning using neural networks interaction [J].
Soheil Mehralian ;
Mohammad Teshnehlab ;
Babak Nasersharif .
Pattern Analysis and Applications, 2021, 24 :1699-1711
[4]   Improving Robustness for Tag Recommendation via Self-Paced Adversarial Metric Learning [J].
Fei, Zhengshun ;
Chen, Jianxin ;
Chen, Gui ;
Xiang, Xinjian .
CMC-COMPUTERS MATERIALS & CONTINUA, 2025, 82 (03) :4237-4261
[5]   Improving Robustness of Hyperbolic Neural Networks by Lipschitz Analysis [J].
Li, Yuekang ;
Mao, Yidan ;
Yang, Yifei ;
Zou, Dongmian .
PROCEEDINGS OF THE 30TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2024, 2024, :1713-1724
[6]   On the Robustness of Metric Learning: An Adversarial Perspective [J].
Huai, Mengdi ;
Zheng, Tianhang ;
Miao, Chenglin ;
Yao, Liuyi ;
Zhang, Aidong .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2022, 16 (05)
[7]   LINEAR DISCRIMINANT ANALYSIS METRIC LEARNING USING SIAMESE NEURAL NETWORKS [J].
Jose, Abin ;
Mei, Qi ;
Eschweiler, Dennis ;
Laube, Ina ;
Stegmaier, Johannes .
2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, :3641-3645
[8]   Improving robustness of convolutional neural networks using element-wise activation scaling [J].
Zhang, Zhi-Yuan ;
Ren, Hao ;
He, Zhenli ;
Zhou, Wei ;
Liu, Di .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2023, 149 :136-148
[9]   Towards Improving Robustness of Deep Neural Networks to Adversarial Perturbations [J].
Amini, Sajjad ;
Ghaemmaghami, Shahrokh .
IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (07) :1889-1903
[10]   A Framework for Metric Learning and Embedding with Topology Learning Neural Networks [J].
Xiang, Zhiyang ;
Xiao, Zhu ;
Wang, Dong .
2015 11TH INTERNATIONAL CONFERENCE ON NATURAL COMPUTATION (ICNC), 2015, :118-122