RMAF: Relu-Memristor-Like Activation Function for Deep Learning

被引:55
|
作者
Yu, Yongbin [1 ]
Adu, Kwabena [1 ]
Tashi, Nyima [2 ]
Anokye, Patrick [1 ]
Wang, Xiangxiang [1 ]
Ayidzoe, Mighty Abra [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Tibet Univ, Sch Informat Sci & Technol, Lhasa 850000, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Biological neural networks; Training; Neurons; Machine learning; Task analysis; Optimization; Activation function; deep learning; memristive window function; muilti-layer perceptron; RMAF; NETWORKS; UNITS;
D O I
10.1109/ACCESS.2020.2987829
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Activation functions facilitate deep neural networks by introducing non-linearity to the learning process. The non-linearity feature gives the neural network the ability to learn complex patterns. Recently, the most widely used activation function is the Rectified Linear Unit (ReLU). Though, other various existing activation including hand-designed alternatives to ReLU have been proposed. However, none has succeeded in replacing ReLU due to their existing inconsistencies. In this work, activation function called ReLU-Memristor-like Activation Function (RMAF) is proposed to leverage benefits of negative values in neural networks. RMAF introduces a constant parameter and a threshold parameter making the function smooth, non-monotonous, and introduces non-linearity in the network. Our experiments show that, the RMAF works better than ReLU and other activation functions on deeper models and across number of challenging datasets. Firstly, experiments are performed by training and classifying on multi-layer perceptron (MLP) over benchmark data such as the Wisconsin breast cancer, MNIST, Iris and Car evaluation. RMAF achieves high performance of 98.74 & x0025;, 99.67 & x0025;, 98.81 & x0025; and 99.42 & x0025; respectively, compared to Sigmoid, Tanh and ReLU. Secondly, experiments were performed on convolution neural network (ResNet) over MNIST, CIFAR-10 and CIFAR-100 data and observed the proposed activation function achieves higher performance accuracy of 99.73 & x0025;, 98.77 & x0025; and 79.82 & x0025; respectively than Tanh, ReLU and Swish. Additionally, we experimented our work on deep networks i.e. squeeze network (SqueezeNet), Dense connected neural network (DenseNet121) and ImageNet dataset, which RMAF produced the best performance. We note that, the RMAF converges faster than the other functions and can replace ReLU in any neural network due to the efficiency, scalability and its similarity to both ReLU and Swish.
引用
收藏
页码:72727 / 72741
页数:15
相关论文
共 50 条
  • [31] PolyLU: A Simple and Robust Polynomial-Based Linear Unit Activation Function for Deep Learning
    Feng, Han-Shen
    Yang, Cheng-Hsiung
    IEEE ACCESS, 2023, 11 : 101347 - 101358
  • [32] Developing Novel Activation Functions Based Deep Learning LSTM for Classification
    Ali, Mohamed H. Essai
    Abdel-Raman, Adel B.
    Badry, Eman A.
    IEEE ACCESS, 2022, 10 : 97259 - 97275
  • [33] Triple-Sigmoid Activation Function for Deep Open-Set Recognition
    Tran, Dinh Tuan
    Shimada, Nobutaka
    Lee, Joo-Ho
    IEEE ACCESS, 2022, 10 : 77668 - 77678
  • [34] PARAMETRIC FLATTEN-T SWISH: AN ADAPTIVE NON-LINEAR ACTIVATION FUNCTION FOR DEEP LEARNING
    Chieng, Hock Hung
    Wahid, Noorhaniza
    Ong, Pauline
    JOURNAL OF INFORMATION AND COMMUNICATION TECHNOLOGY-MALAYSIA, 2021, 20 (01): : 21 - 39
  • [35] ErfReLU: adaptive activation function for deep neural network
    Rajanand, Ashish
    Singh, Pradeep
    PATTERN ANALYSIS AND APPLICATIONS, 2024, 27 (02)
  • [36] Proposal of New Activation Function in Deep Image Prior
    Segawa, Ryo
    Hayashi, Hitoshi
    Fujii, Shohei
    IEEJ TRANSACTIONS ON ELECTRICAL AND ELECTRONIC ENGINEERING, 2020, 15 (08) : 1248 - 1249
  • [37] Deep-Learning Software Reliability Model Using SRGM as Activation Function
    Kim, Youn Su
    Pham, Hoang
    Chang, In Hong
    APPLIED SCIENCES-BASEL, 2023, 13 (19):
  • [38] Protein function prediction with gene ontology: from traditional to deep learning models
    Thi Thuy Duong Vu
    Jung, Jaehee
    PEERJ, 2021, 9
  • [39] Neural Network Physically Unclonable Function: A Trainable Physically Unclonable Function System with Unassailability against Deep Learning Attacks Using Memristor Array
    Park, Junkyu
    Lee, Yoonji
    Jeong, Hakcheon
    Choi, Shinhyun
    ADVANCED INTELLIGENT SYSTEMS, 2021, 3 (11)
  • [40] Generalization of Deep Learning Gesture Classification in Robotic-Assisted Surgical Data: From Dry Lab to Clinical-Like Data
    Itzkovich, Danit
    Sharon, Yarden
    Jarc, Anthony
    Refaely, Yael
    Nisky, Ilana
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (03) : 1329 - 1340