RMAF: Relu-Memristor-Like Activation Function for Deep Learning

被引:55
|
作者
Yu, Yongbin [1 ]
Adu, Kwabena [1 ]
Tashi, Nyima [2 ]
Anokye, Patrick [1 ]
Wang, Xiangxiang [1 ]
Ayidzoe, Mighty Abra [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu 610054, Peoples R China
[2] Tibet Univ, Sch Informat Sci & Technol, Lhasa 850000, Peoples R China
来源
IEEE ACCESS | 2020年 / 8卷
基金
中国国家自然科学基金;
关键词
Biological neural networks; Training; Neurons; Machine learning; Task analysis; Optimization; Activation function; deep learning; memristive window function; muilti-layer perceptron; RMAF; NETWORKS; UNITS;
D O I
10.1109/ACCESS.2020.2987829
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Activation functions facilitate deep neural networks by introducing non-linearity to the learning process. The non-linearity feature gives the neural network the ability to learn complex patterns. Recently, the most widely used activation function is the Rectified Linear Unit (ReLU). Though, other various existing activation including hand-designed alternatives to ReLU have been proposed. However, none has succeeded in replacing ReLU due to their existing inconsistencies. In this work, activation function called ReLU-Memristor-like Activation Function (RMAF) is proposed to leverage benefits of negative values in neural networks. RMAF introduces a constant parameter and a threshold parameter making the function smooth, non-monotonous, and introduces non-linearity in the network. Our experiments show that, the RMAF works better than ReLU and other activation functions on deeper models and across number of challenging datasets. Firstly, experiments are performed by training and classifying on multi-layer perceptron (MLP) over benchmark data such as the Wisconsin breast cancer, MNIST, Iris and Car evaluation. RMAF achieves high performance of 98.74 & x0025;, 99.67 & x0025;, 98.81 & x0025; and 99.42 & x0025; respectively, compared to Sigmoid, Tanh and ReLU. Secondly, experiments were performed on convolution neural network (ResNet) over MNIST, CIFAR-10 and CIFAR-100 data and observed the proposed activation function achieves higher performance accuracy of 99.73 & x0025;, 98.77 & x0025; and 79.82 & x0025; respectively than Tanh, ReLU and Swish. Additionally, we experimented our work on deep networks i.e. squeeze network (SqueezeNet), Dense connected neural network (DenseNet121) and ImageNet dataset, which RMAF produced the best performance. We note that, the RMAF converges faster than the other functions and can replace ReLU in any neural network due to the efficiency, scalability and its similarity to both ReLU and Swish.
引用
收藏
页码:72727 / 72741
页数:15
相关论文
共 50 条
  • [41] Recursion Newton-Like Algorithm for l2,0-ReLU Deep Neural Networks
    Zhang, Hui
    Yuan, Zhengpeng
    Xiu, Naihua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 5882 - 5896
  • [42] Fast Activation Function Approach for Deep Learning Based Online Anomaly Intrusion Detection
    Alrawashdeh, Khaled
    Purdy, Carla
    2018 IEEE 4TH INTERNATIONAL CONFERENCE ON BIG DATA SECURITY ON CLOUD (BIGDATASECURITY), 4THIEEE INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE AND SMART COMPUTING, (HPSC) AND 3RD IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT DATA AND SECURITY (IDS), 2018, : 5 - 13
  • [43] The Negative BER Loss Function for Deep Learning Decoders
    Dong, Rui
    Lu, Fang
    Dong, Yan
    Yan, Haotian
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (08) : 1824 - 1828
  • [44] A method for composite activation functions in deep learning for object detection
    Liao, Jing
    Yu, Chang
    Jiang, Lei
    Guo, Linpei
    Liang, Wei
    Li, Kuanching
    Pathan, Al-Sakib Khan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (05)
  • [45] THE STUDY OF ACTIVATION FUNCTIONS IN DEEP LEARNING FOR PEDESTRIAN DETECTION AND TRACKING
    Favorskaya, M. N.
    Andreev, V. V.
    INTERNATIONAL WORKSHOP ON PHOTOGRAMMETRIC AND COMPUTER VISION TECHNIQUES FOR VIDEO SURVEILLANCE, BIOMETRICS AND BIOMEDICINE, 2019, 42-2 (W12): : 53 - 59
  • [46] SieveNet: Decoupling activation function neural network for privacy-preserving deep learning
    Wang, Qizheng
    Ma, Wenping
    Liu, Ge
    INFORMATION SCIENCES, 2021, 573 : 262 - 278
  • [47] RELU-LIKE NON-MONOTONIC SMOOTH ACTIVATION FUNCTIONS BASED ON REGULARIZED HEAVISIDE FUNCTIONS AND EXTENSIONS
    Ren, Pengfei
    Pan, Tony yuxiang
    Yang, Guangyu
    Guo, Yanchen
    Wei, Weibo
    Pan, Zhenkuan
    MATHEMATICAL FOUNDATIONS OF COMPUTING, 2025,
  • [48] Diverse Activation Functions in Deep Learning
    Wang, Bin
    Li, Tianrui
    Huang, Yanyong
    Luo, Huaishao
    Guo, Dongming
    Horng, Shi-Jinn
    2017 12TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND KNOWLEDGE ENGINEERING (IEEE ISKE), 2017,
  • [49] A Comprehensive Survey of Deep Learning Techniques in Protein Function Prediction
    Dhanuka, Richa
    Singh, Jyoti Prakash
    Tripathi, Anushree
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2023, 20 (03) : 2291 - 2301
  • [50] A novel loss function of deep learning in wind speed forecasting
    Chen, Xi
    Yu, Ruyi
    Ullah, Sajid
    Wu, Dianming
    Li, Zhiqiang
    Li, Qingli
    Qi, Honggang
    Liu, Jihui
    Liu, Min
    Zhang, Yundong
    ENERGY, 2022, 238