Generalized Correntropy based deep learning in presence of non-Gaussian noises

被引:28
作者
Chen, Liangjun [1 ]
Qu, Hua [1 ]
Zhao, Jihong [1 ,2 ]
机构
[1] Xi An Jiao Tong Univ, Sch Elect & Informat Engn, Xian 710049, Shaanxi, Peoples R China
[2] Xian Univ Posts & Telecommun, Sch Telecommun & Informat Engn, Xian 710061, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Generalized Correntropy; Stacked autoencoders; Non-Gaussian noise; Network traffic classification; MAXIMUM; ALGORITHM;
D O I
10.1016/j.neucom.2017.06.080
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning algorithms are the hottest topics in machine learning area lately. Although deep learning has made great progress in many domains, the robustness of learning systems with deep architectures is still rarely studied and needs further investigation. For instance, the impulsive noises (or outliers) are pervasive in real world data and can badly influence the mean square error (MSE) cost function based deep learning algorithms. Correntropy based loss function, which uses Gaussian kernel, is widely utilized to reject the above noises, however, the effect is not satisfactory. Therefore, generalized Correntropy (GC) is put forward to further improve the robustness, which uses generalized Gaussian density (GGD) function as kernel. GC can achieve extra flexibility through the GC parameters, which control the behavior of the induced metric, and shows a markedly better robustness than Correntropy. Motivated by the enhanced robustness of GC, we propose a new robust algorithm named generalized Correntropy based stacked autoencoder (GC-SAE), which is developed by combining the GC and stacked autoencoder (SAE). The new algorithms can extract useful features from the data corrupted by impulsive noises (or outliers) in a more effective way. The good robustness of the proposed method is confirmed by the experimental results on MNIST benchmark dataset. Furthermore, we show how our model can be applied for robust network classification, based on Moore network data of 377,526 samples with 12 classes. (C) 2017 Elsevier B. V. All rights reserved.
引用
收藏
页码:41 / 50
页数:10
相关论文
共 48 条
  • [1] [Anonymous], 2006, Proc. Adv. Neural Inf. Process. Syst.
  • [2] [Anonymous], 2007, IEEE INT C ICML
  • [3] Bayesian neural networks for Internet traffic classification
    Auld, Tom
    Moore, Andrew W.
    Gull, Stephen F.
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2007, 18 (01): : 223 - 239
  • [4] Bengio Y., 2006, ADV NEURAL INFORM PR, V19
  • [5] Learning Deep Architectures for AI
    Bengio, Yoshua
    [J]. FOUNDATIONS AND TRENDS IN MACHINE LEARNING, 2009, 2 (01): : 1 - 127
  • [6] Maximum correntropy Kalman filter
    Chen, Badong
    Liu, Xi
    Zhao, Haiquan
    Principe, Jose C.
    [J]. AUTOMATICA, 2017, 76 : 70 - 77
  • [7] Generalized Correntropy for Robust Adaptive Filtering
    Chen, Badong
    Xing, Lei
    Zhao, Haiquan
    Zheng, Nanning
    Principe, Jose C.
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (13) : 3376 - 3387
  • [8] Convergence of a Fixed-Point Algorithm under Maximum Correntropy Criterion
    Chen, Badong
    Wang, Jianji
    Zhao, Haiquan
    Zheng, Nanning
    Principe, Jose C.
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2015, 22 (10) : 1723 - 1727
  • [9] Steady-State Mean-Square Error Analysis for Adaptive Filtering under the Maximum Correntropy Criterion
    Chen, Badong
    Xing, Lei
    Liang, Junli
    Zheng, Nanning
    Principe, Jose C.
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2014, 21 (07) : 880 - 884
  • [10] Maximum Correntropy Estimation Is a Smoothed MAP Estimation
    Chen, Badong
    Principe, Jose C.
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2012, 19 (08) : 491 - 494