Generalized Adversarial Training in Riemannian Space

被引:6
作者
Zhang, Shufei [1 ]
Huang, Kaizhu [1 ]
Zhang, Rui [2 ]
Hussain, Amir [3 ]
机构
[1] Xian Jiaotong Liverpool Univ, Dept Elect & Elect Engn, Suzhou, Peoples R China
[2] Xian Jiaotong Liverpool Univ, Dept Math Sci, Suzhou, Peoples R China
[3] Edinburgh Napier Univ, Sch Comp, Edinburgh, Midlothian, Scotland
来源
2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019) | 2019年
关键词
D O I
10.1109/ICDM.2019.00093
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adversarial examples, referred to as augmented data points generated by imperceptible perturbations of input samples, have recently drawn much attention. Well-crafted adversarial examples may even mislead state-of-the-art deep neural network (DNN) models to make wrong predictions easily. To alleviate this problem, many studies have focused on investigating how adversarial examples can be generated and/or effectively handled. All existing works tackle this problem in the Euclidean space. In this paper, we extend the learning of adversarial examples to the more general Riemannian space over DNNs. The proposed work is important in that (1) it is a generalized learning methodology since Riemmanian space will be degraded to the Euclidean space in a special case; (2) it is the first work to tackle the adversarial example problem tractably through the perspective of Riemannian geometry; (3) from the perspective of geometry, our method leads to the steepest direction of the loss function, by considering the second order information of the loss function. We also provide a theoretical study showing that our proposed method can truly find the descent direction for the loss function, with a comparable computational time against traditional adversarial methods. Finally, the proposed framework demonstrates superior performance over traditional counterpart methods, using benchmark data including MNIST, CIFAR-10 and SVHN.
引用
收藏
页码:826 / 835
页数:10
相关论文
共 27 条
  • [1] Natural gradient works efficiently in learning
    Amari, S
    [J]. NEURAL COMPUTATION, 1998, 10 (02) : 251 - 276
  • [2] Nguyen A, 2015, PROC CVPR IEEE, P427, DOI 10.1109/CVPR.2015.7298640
  • [3] [Anonymous], 2015, ARXIV151105432
  • [4] AN INVERSE MATRIX ADJUSTMENT ARISING IN DISCRIMINANT ANALYSIS
    BARTLETT, MS
    [J]. ANNALS OF MATHEMATICAL STATISTICS, 1951, 22 (01): : 107 - 111
  • [5] Biggio B., 2013, MACHINE LEARNING KNO, P387, DOI [DOI 10.1007/978-3-642-40994, DOI 10.1007/978-3-642-40994-3_25]
  • [6] REPRESENTATIONS OF QUASI-NEWTON MATRICES AND THEIR USE IN LIMITED MEMORY METHODS
    BYRD, RH
    NOCEDAL, J
    SCHNABEL, RB
    [J]. MATHEMATICAL PROGRAMMING, 1994, 63 (02) : 129 - 156
  • [7] Dauphin YN, 2014, ADV NEUR IN, V27
  • [8] Goodfellow I J, 2015, P INT C LEARN REPR I
  • [9] Identity Mappings in Deep Residual Networks
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 : 630 - 645
  • [10] Densely Connected Convolutional Networks
    Huang, Gao
    Liu, Zhuang
    van der Maaten, Laurens
    Weinberger, Kilian Q.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2261 - 2269