Enhance the Performance of Deep Neural Networks via L2 Regularization on the Input of Activations

被引:0
|
作者
Guang Shi
Jiangshe Zhang
Huirong Li
Changpeng Wang
机构
[1] Xi’an Jiaotong University,School of Mathematics and Statistics
[2] Shangluo University,Department of Mathematics and Computer Application
[3] Chang’an University,School of Mathematics and Information Science
来源
Neural Processing Letters | 2019年 / 50卷
关键词
Neural networks; ReLU; Saturation phenomenon; L2 regularization;
D O I
暂无
中图分类号
学科分类号
摘要
Deep neural networks (DNNs) are witnessing increasing attention in machine learning. However, the information propagation is becoming increasingly difficult as the networks get deeper, which makes the optimization of DNN extremely hard. One reason of this difficulty is saturation of hidden units. In this paper, we propose a novel methodology named RegA to decrease the influences of saturation on ReLU-DNNs (DNNs with ReLU). Instead of changing the activation functions or the initialization strategy, our methodology explicitly encourage the pre-activation to be out of the saturation region. Specifically, we add an auxiliary objective induced by L2-norm of the pre-activation values to the optimization problem. The auxiliary objective could help to active more units and promote effective information propagation in ReLU-DNNs. By conducting experiments on several large-scale real datasets, we demonstrate better representations could be learned by using RegA and the method help ReLU-DNNs get better performance on convergence and accuracy.
引用
收藏
页码:57 / 75
页数:18
相关论文
共 50 条
  • [31] Sparse synthesis regularization with deep neural networks
    Obmann, Daniel
    Schwab, Johannes
    Haltmeier, Markus
    2019 13TH INTERNATIONAL CONFERENCE ON SAMPLING THEORY AND APPLICATIONS (SAMPTA), 2019,
  • [32] Group sparse regularization for deep neural networks
    Scardapane, Simone
    Comminiello, Danilo
    Hussain, Amir
    Uncini, Aurelio
    NEUROCOMPUTING, 2017, 241 : 81 - 89
  • [33] Sentiment Analysis of Tweets by Convolution Neural Network with L1 and L2 Regularization
    Rangra, Abhilasha
    Sehgal, Vivek Kumar
    Shukla, Shailendra
    ADVANCED INFORMATICS FOR COMPUTING RESEARCH, ICAICR 2018, PT I, 2019, 955 : 355 - 365
  • [34] LocalDrop: A Hybrid Regularization for Deep Neural Networks
    Lu, Ziqing
    Xu, Chang
    Du, Bo
    Ishida, Takashi
    Zhang, Lefei
    Sugiyama, Masashi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) : 3590 - 3601
  • [35] A Comparison of Regularization Techniques in Deep Neural Networks
    Nusrat, Ismoilov
    Jang, Sung-Bong
    SYMMETRY-BASEL, 2018, 10 (11):
  • [36] Structured Pruning of Convolutional Neural Networks via L1 Regularization
    Yang, Chen
    Yang, Zhenghong
    Khattak, Abdul Mateen
    Yang, Liu
    Zhang, Wenxin
    Gao, Wanlin
    Wang, Minjuan
    IEEE ACCESS, 2019, 7 : 106385 - 106394
  • [37] A novel framework to enhance the performance of training distributed deep neural networks
    Phan, Trung
    Do, Phuc
    INTELLIGENT DATA ANALYSIS, 2023, 27 (03) : 753 - 768
  • [38] Training Deep Photonic Convolutional Neural Networks With Sinusoidal Activations
    Passalis, Nikolaos
    Mourgias-Alexandris, George
    Tsakyridis, Apostolos
    Pleros, Nikos
    Tefas, Anastasios
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2021, 5 (03): : 384 - 393
  • [39] Scaling Deep Spiking Neural Networks with Binary Stochastic Activations
    Roy, Deboleena
    Chakraborty, Indranil
    Roy, Kaushik
    2019 IEEE INTERNATIONAL CONFERENCE ON COGNITIVE COMPUTING (IEEE ICCC 2019), 2019, : 50 - 58
  • [40] Deep Neural Networks With Trainable Activations and Controlled Lipschitz Constant
    Aziznejad, Shayan
    Gupta, Harshit
    Campos, Joaquim
    Unser, Michael
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2020, 68 : 4688 - 4699