RMAF: Relu-Memristor-Like Activation Function for Deep Learning

被引:55
|
作者
Yu, Yongbin [1 ]
Adu, Kwabena [1 ]
Tashi, Nyima [2 ]
Anokye, Patrick [1 ]
Wang, Xiangxiang [1 ]
Ayidzoe, Mighty Abra [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Tibet Univ, Sch Informat Sci & Technol, Lhasa 850000, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Biological neural networks; Training; Neurons; Machine learning; Task analysis; Optimization; Activation function; deep learning; memristive window function; muilti-layer perceptron; RMAF; NETWORKS; UNITS;
D O I
10.1109/ACCESS.2020.2987829
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Activation functions facilitate deep neural networks by introducing non-linearity to the learning process. The non-linearity feature gives the neural network the ability to learn complex patterns. Recently, the most widely used activation function is the Rectified Linear Unit (ReLU). Though, other various existing activation including hand-designed alternatives to ReLU have been proposed. However, none has succeeded in replacing ReLU due to their existing inconsistencies. In this work, activation function called ReLU-Memristor-like Activation Function (RMAF) is proposed to leverage benefits of negative values in neural networks. RMAF introduces a constant parameter and a threshold parameter making the function smooth, non-monotonous, and introduces non-linearity in the network. Our experiments show that, the RMAF works better than ReLU and other activation functions on deeper models and across number of challenging datasets. Firstly, experiments are performed by training and classifying on multi-layer perceptron (MLP) over benchmark data such as the Wisconsin breast cancer, MNIST, Iris and Car evaluation. RMAF achieves high performance of 98.74 & x0025;, 99.67 & x0025;, 98.81 & x0025; and 99.42 & x0025; respectively, compared to Sigmoid, Tanh and ReLU. Secondly, experiments were performed on convolution neural network (ResNet) over MNIST, CIFAR-10 and CIFAR-100 data and observed the proposed activation function achieves higher performance accuracy of 99.73 & x0025;, 98.77 & x0025; and 79.82 & x0025; respectively than Tanh, ReLU and Swish. Additionally, we experimented our work on deep networks i.e. squeeze network (SqueezeNet), Dense connected neural network (DenseNet121) and ImageNet dataset, which RMAF produced the best performance. We note that, the RMAF converges faster than the other functions and can replace ReLU in any neural network due to the efficiency, scalability and its similarity to both ReLU and Swish.
引用
收藏
页码:72727 / 72741
页数:15
相关论文
共 50 条
  • [21] A leading adaptive activation function for deep reinforcement learning
    Jia, Chongjiexin
    Li, Tuanjie
    Dong, Hangjia
    Xie, Chao
    Peng, Wenxuan
    Ning, Yuming
    JOURNAL OF COMPUTATIONAL SCIENCE, 2025, 88
  • [22] Self-Growing Binary Activation Network: A Novel Deep Learning Model With Dynamic Architecture
    Zhang, Zeyang
    Chen, Yidong
    Zhou, Changle
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (01) : 624 - 633
  • [23] Universal Function Approximation Through Over-the-Air Computing: A Deep Learning Approach
    Bouzinis, Pavlos S.
    Evgenidis, Nikos G.
    Mitsiou, Nikos A.
    Tegos, Sotiris A.
    Diamantoulakis, Panagiotis D.
    Karagiannidis, George K.
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2024, 5 : 2958 - 2967
  • [24] HcLSH: A Novel Non-Linear Monotonic Activation Function for Deep Learning Methods
    Abdel-Nabi, Heba
    Al-Naymat, Ghazi
    Ali, Mostafa Z.
    Awajan, Arafat
    IEEE ACCESS, 2023, 11 : 47794 - 47815
  • [25] Parametric RSigELU: a new trainable activation function for deep learning
    Kilicarslan, Serhat
    Celik, Mete
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (13) : 7595 - 7607
  • [26] Enhancing Deep Learning Models for Brain Tumor Detection: The Impact of Activation Function Selection
    Nasayreh, Ahmad
    Jaradat, Ameera S.
    Al Mamlook, Rabia Emhamed
    Bashkami, Ayah
    Gharaibeh, Hasan
    Al-Na'amneh, Qais
    Bzizi, Hanin
    4TH INTERDISCIPLINARY CONFERENCE ON ELECTRICS AND COMPUTER, INTCEC 2024, 2024,
  • [27] Hardware Implementation of a Softmax-Like Function for Deep Learning†
    Kouretas, Ioannis
    Paliouras, Vassilis
    TECHNOLOGIES, 2020, 8 (03)
  • [28] Learnable Leaky ReLU (LeLeLU): An Alternative Accuracy-Optimized Activation Function
    Maniatopoulos, Andreas
    Mitianoudis, Nikolaos
    INFORMATION, 2021, 12 (12)
  • [29] A Detection Method Against Selfish Mining-Like Attacks Based on Ensemble Deep Learning in IoT
    Wang, Yilei
    Li, Chunmei
    Zhang, Yiting
    Li, Tao
    Ning, Jianting
    Gai, Keke
    Choo, Kim-Kwang Raymond
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (11): : 19564 - 19574
  • [30] Universal Function Approximation by Deep Neural Nets with Bounded Width and ReLU Activations
    Hanin, Boris
    MATHEMATICS, 2019, 7 (10)