Enhancement of neural networks with an alternative activation function tanhLU

被引:65
作者
Shen, Shui-Long [1 ]
Zhang, Ning [1 ,2 ]
Zhou, Annan [3 ]
Yin, Zhen-Yu [2 ]
机构
[1] Shantou Univ, Coll Engn, Dept Civil & Environm Engn, MOE Key Lab Intelligence Mfg Technol, Shantou 515063, Guangdong, Peoples R China
[2] Hong Kong Polytech Univ, Dept Civil & Environm Engn, Hung Hom, Kowloon, Hong Kong, Peoples R China
[3] RMIT Univ, Sch Engn, Melbourne, Vic 3001, Australia
关键词
Neural networks; Activation function; tanhLUs;
D O I
10.1016/j.eswa.2022.117181
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A novel activation function (referred to as tanhLU) that integrates hyperbolic tangent function (tanh) with a linear unit is proposed as a promising alternative to tanh for neural networks. The tanhLU is inspired by the boundlessness of rectified linear unit (ReLU) and the symmetry of tanh. Three variable parameters in tanhLU controlling activation values and gradients could be preconfigured as constants or adaptively optimized during the training process. The capacity of tanhLU is first investigated by checking the weight gradients in error back propagation. Experiments are conducted to validate the improvement of tanhLUs on five types of neural networks, based on seven benchmark datasets in different domains. tanhLU is then applied to predict the highly nonlinear stress-strain relationship of soils by using the multiscale stress-strain (MSS) dataset. The experiment results indicate that using constant tanhLU leads to apparent improvement on FCNN and LSTM with lower loss and higher accuracy compared with tanh. Adaptive tanhLUs achieved the state-of-the-art performance for multiple deep neural networks in image classification and face recognition.
引用
收藏
页数:13
相关论文
共 47 条
  • [1] Agostinelli F., 2015, ARXIV14126830V3
  • [2] [Anonymous], P 30 INT C MACH LEAR
  • [3] [Anonymous], 2017, ARXIV170301619
  • [4] [Anonymous], 2015, RECURRENT NEURAL NET
  • [5] SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation
    Badrinarayanan, Vijay
    Kendall, Alex
    Cipolla, Roberto
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2481 - 2495
  • [6] Bai S, 2019, ARXIV
  • [7] Linearized sigmoidal activation: A novel activation function with tractable non-linear characteristics to boost representation capability
    Bawa, Vivek Singh
    Kumar, Vinay
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 120 : 346 - 356
  • [8] Bergstra J., 2009, Quadratic polynomials learn better image features
  • [9] Brown P. F., 1992, Computational Linguistics, V18, P31
  • [10] Cho K., 2014, P C EMP METH NAT LAN, P1724