RMAF: Relu-Memristor-Like Activation Function for Deep Learning

被引:55
|
作者
Yu, Yongbin [1 ]
Adu, Kwabena [1 ]
Tashi, Nyima [2 ]
Anokye, Patrick [1 ]
Wang, Xiangxiang [1 ]
Ayidzoe, Mighty Abra [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Tibet Univ, Sch Informat Sci & Technol, Lhasa 850000, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Biological neural networks; Training; Neurons; Machine learning; Task analysis; Optimization; Activation function; deep learning; memristive window function; muilti-layer perceptron; RMAF; NETWORKS; UNITS;
D O I
10.1109/ACCESS.2020.2987829
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Activation functions facilitate deep neural networks by introducing non-linearity to the learning process. The non-linearity feature gives the neural network the ability to learn complex patterns. Recently, the most widely used activation function is the Rectified Linear Unit (ReLU). Though, other various existing activation including hand-designed alternatives to ReLU have been proposed. However, none has succeeded in replacing ReLU due to their existing inconsistencies. In this work, activation function called ReLU-Memristor-like Activation Function (RMAF) is proposed to leverage benefits of negative values in neural networks. RMAF introduces a constant parameter and a threshold parameter making the function smooth, non-monotonous, and introduces non-linearity in the network. Our experiments show that, the RMAF works better than ReLU and other activation functions on deeper models and across number of challenging datasets. Firstly, experiments are performed by training and classifying on multi-layer perceptron (MLP) over benchmark data such as the Wisconsin breast cancer, MNIST, Iris and Car evaluation. RMAF achieves high performance of 98.74 & x0025;, 99.67 & x0025;, 98.81 & x0025; and 99.42 & x0025; respectively, compared to Sigmoid, Tanh and ReLU. Secondly, experiments were performed on convolution neural network (ResNet) over MNIST, CIFAR-10 and CIFAR-100 data and observed the proposed activation function achieves higher performance accuracy of 99.73 & x0025;, 98.77 & x0025; and 79.82 & x0025; respectively than Tanh, ReLU and Swish. Additionally, we experimented our work on deep networks i.e. squeeze network (SqueezeNet), Dense connected neural network (DenseNet121) and ImageNet dataset, which RMAF produced the best performance. We note that, the RMAF converges faster than the other functions and can replace ReLU in any neural network due to the efficiency, scalability and its similarity to both ReLU and Swish.
引用
收藏
页码:72727 / 72741
页数:15
相关论文
共 50 条
  • [1] Optimization of ReLU Activation Function for Deep-Learning-based Breast Cancer Classification on Mammogram Images
    Razali, Noor Fadzilah
    Isa, Iza Sazanita
    Sulaiman, Siti Noraini
    Osman, Muhammad Khusairi
    Karim, Noor Khairiah A.
    Nordin, Siti Aminah
    2024 IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC CONTROL AND INTELLIGENT SYSTEMS, I2CACIS 2024, 2024, : 267 - 272
  • [2] Effects of the Flatness Network Parameter Threshold on the Performance of the Rectified Linear Unit Memristor-Like Activation Function in Deep Learning
    Marcelle Tchepgoua Mbakop
    Justin Roger Mboupda Pone
    Priva Chassem Kamdem
    Romanic Kengne
    SN Computer Science, 5 (8)
  • [3] Construction of Deep ReLU Nets for Spatially Sparse Learning
    Liu, Xia
    Wang, Di
    Lin, Shao-Bo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (10) : 7746 - 7760
  • [4] TeLU: A New Activation Function for Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    2020 14TH INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND TELECOMMUNICATIONS (ISETC), 2020, : 32 - 35
  • [5] A Universal Activation Function for Deep Learning
    Hwang, Seung-Yeon
    Kim, Jeong-Joon
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 75 (02): : 3553 - 3569
  • [6] Deep Learning-Based Beamforming for Millimeter-Wave Systems Using Parametric ReLU Activation Function
    Alshimaa H. Ismail
    Tarek Abed Soliman
    Mohamed Rihan
    Moawad I. Dessouky
    Wireless Personal Communications, 2023, 129 : 825 - 836
  • [7] Deep Learning-Based Beamforming for Millimeter-Wave Systems Using Parametric ReLU Activation Function
    Ismail, Alshimaa H. H.
    Soliman, Tarek Abed
    Rihan, Mohamed
    Dessouky, Moawad I. I.
    WIRELESS PERSONAL COMMUNICATIONS, 2023, 129 (02) : 825 - 836
  • [8] All-optical ultrafast ReLU function for energy-efficient nanophotonic deep learning
    Li, Gordon H. Y.
    Sekine, Ryoto
    Nehra, Rajveer
    Gray, Robert M.
    Ledezma, Luis
    Guo, Qiushi
    Marandi, Alireza
    NANOPHOTONICS, 2023, 12 (05) : 847 - 855
  • [9] P-Swish: Activation Function with Learnable Parameters Based on Swish Activation Function in Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    2020 14TH INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND TELECOMMUNICATIONS (ISETC), 2020, : 40 - 43
  • [10] MorphoActivation: Generalizing ReLU Activation Function by Mathematical Morphology
    Velasco-Forero, Santiago
    Angulo, Jesus
    DISCRETE GEOMETRY AND MATHEMATICAL MORPHOLOGY, DGMM 2022, 2022, 13493 : 449 - 461