P plus FELU: Flexible and trainable fast exponential linear unit for deep learning architectures

被引:5
作者
Adem, Kemal [1 ]
机构
[1] Sivas Univ Sci & Technol, Dept Comp Engn, Sivas, Turkey
关键词
Activation function; Deep learning architectures; Flexible and trainable P plus FELU; ACTIVATION FUNCTION;
D O I
10.1007/s00521-022-07625-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Activation functions have an important role in obtaining the most appropriate output by processing the information coming into the network in deep learning architectures. Deep learning architectures are widely used in areas such as image processing applications, time series, and disease classification, generally in line with the analysis of large and complex data. Choosing the appropriate architecture and activation function is an important factor in achieving successful learning and classification performance. There are many studies to improve the performance of deep learning architectures and to overcome the disappearing gradient and negative region problems in activation functions. A flexible and trainable fast exponential linear unit (P + FELU) activation function is proposed to overcome existing problems. With the proposed P + FELU activation function, a higher success rate and faster calculation time can be achieved by incorporating the advantages of fast exponentially linear unit (FELU), exponential linear unit (ELU), and rectified linear unit (RELU) activation functions. Performance evaluations of the proposed P + FELU activation function were made on MNIST, CIFAR-10, and CIFAR-100 benchmark datasets. Experimental evaluations have shown that the proposed activation function outperforms the ReLU, ELU, SELU, MPELU, TReLU, and FELU activation functions and effectively improves the noise robustness of the network. Experimental results show that this activation function with "flexible and trainable" properties can effectively prevent vanishing gradient and make multilayer perceptron neural networks deeper.
引用
收藏
页码:21729 / 21740
页数:12
相关论文
共 33 条
[1]   Defect detection of seals in multilayer aseptic packages using deep learning [J].
Adem, Kemal ;
Kozkurt, Cemil .
TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2019, 27 (06) :4220-4230
[2]   Exudate detection for diabetic retinopathy with circular Hough transformation and convolutional neural networks [J].
Adem, Kemal .
EXPERT SYSTEMS WITH APPLICATIONS, 2018, 114 :289-295
[3]   Linearized sigmoidal activation: A novel activation function with tractable non-linear characteristics to boost representation capability [J].
Bawa, Vivek Singh ;
Kumar, Vinay .
EXPERT SYSTEMS WITH APPLICATIONS, 2019, 120 :346-356
[4]  
Clevert D.-A., 2016, C TRACK P, P1
[5]   The Deep Features and Attention Mechanism-Based Method to Dish Healthcare Under Social IoT Systems: An Empirical Study With a Hand-Deep Local-Global Net [J].
Gao, Honghao ;
Xu, Kaili ;
Cao, Min ;
Xiao, Junsheng ;
Xu, Qiang ;
Yin, Yuyu .
IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2022, 9 (01) :336-347
[6]  
Glorot X., 2010, P 13 INT C ART INT S, P249
[7]  
Godfrey LB, 2019, IEEE SYS MAN CYBERN, P3006, DOI [10.1109/smc.2019.8913972, 10.1109/SMC.2019.8913972]
[8]   Dual Rectified Linear Units (DReLUs): A replacement for tanh activation functions in Quasi-Recurrent Neural Networks [J].
Godin, Frederic ;
Degrave, Jonas ;
Dambre, Joni ;
De Neye, Wesley .
PATTERN RECOGNITION LETTERS, 2018, 116 :8-14
[9]  
Gupta SF, 2017, 2017 IEEE INTERNATIONAL CONFERENCE ON ADVANCED NETWORKS AND TELECOMMUNICATIONS SYSTEMS (ANTS)
[10]   Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1026-1034