Activation Functions and Their Characteristics in Deep Neural Networks

被引:0
|
作者
Ding, Bin [1 ]
Qian, Huimin [1 ]
Zhou, Jun [1 ]
机构
[1] Hohai Univ, Coll Energy & Elect Engn, Nanjing 211100, Peoples R China
来源
PROCEEDINGS OF THE 30TH CHINESE CONTROL AND DECISION CONFERENCE (2018 CCDC) | 2018年
关键词
neural network; deep architecture; activation function;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks have gained remarkable achievements in many research areas, especially in computer vision, and natural language processing. The great successes of deep neural networks depend on several aspects in which the development of activation function is one of the most important elements. Being aware of this, a number of researches have concentrated on the performance improvements after the revision of a certain activation function in some specified neural networks. We have noticed that there are few papers to review thoroughly the activation functions employed by the neural networks. Therefore, considering the impact of improving the performance of neural networks with deep architectures, the status and the developments of commonly used activation functions will be investigated in this paper. More specifically, the definitions, the impacts on the neural networks, and the advantages and disadvantages of quite a few activation functions will be discussed in this paper. Furthermore, experimental results on the dataset MNIST are employed to compare the performance of different activation functions.
引用
收藏
页码:1836 / 1841
页数:6
相关论文
共 50 条
  • [21] Derivation of NARX models by expanding activation functions in neural networks
    Inaoka, Hidenori
    Kobayashi, Kozue
    Nebuya, Satoru
    Kumagai, Hiroshi
    Tsuruta, Harukazu
    Fukuoka, Yutaka
    IEEJ TRANSACTIONS ON ELECTRICAL AND ELECTRONIC ENGINEERING, 2019, 14 (08) : 1209 - 1218
  • [22] Bayesian Optimization for Sparse Neural Networks With Trainable Activation Functions
    Fakhfakh, Mohamed
    Chaari, Lotfi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (10) : 6699 - 6712
  • [23] Application of neural networks with orthogonal activation functions in control of dynamical systems
    Nikolic, Sasa S.
    Antic, Dragan S.
    Milojkovic, Marko T.
    Milovanovic, Miroslav B.
    Peric, Stanisa Lj.
    Mitic, Darko B.
    INTERNATIONAL JOURNAL OF ELECTRONICS, 2016, 103 (04) : 667 - 685
  • [24] Phase transition analysis for shallow neural networks with arbitrary activation functions
    Citton, Otavio
    Richert, Frederieke
    Biehl, Michael
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2025, 660
  • [25] Heuristic Search for Activation Functions of Neural Networks Based on Gaussian Processes
    Shi, Xinxing
    Chen, Jialin
    Wang, Lingli
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [26] GAAF: Searching Activation Functions for Binary Neural Networks Through Genetic Algorithm
    Li, Yanfei
    Geng, Tong
    Stein, Samuel
    Li, Ang
    Yu, Huimin
    TSINGHUA SCIENCE AND TECHNOLOGY, 2023, 28 (01): : 207 - 220
  • [27] GPU-based Empirical Evaluation of Activation Functions in Convolutional Neural Networks
    Zaheer, Raniah
    Shaziya, Humera
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON INVENTIVE SYSTEMS AND CONTROL (ICISC 2018), 2018, : 769 - 773
  • [28] Stability Analysis for a Class of Delayed Neural Networks with Nonlinear Homogeneous Activation Functions
    Wang, Man
    Chen, Boshan
    2015 SIXTH INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL AND INFORMATION PROCESSING (ICICIP), 2015, : 30 - 35
  • [29] Periodic Function as Activation Function for Neural Networks
    Xu, Ding
    Guan, Yue
    Cai, Ping-ping
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE: TECHNIQUES AND APPLICATIONS, AITA 2016, 2016, : 179 - 183
  • [30] Accelerating the Activation Function Selection for Hybrid Deep Neural Networks - FPGA Implementation
    Waseem, Shaik Mohammed
    Suraj, Alavala Venkata
    Roy, Subir Kumar
    2021 IEEE REGION 10 SYMPOSIUM (TENSYMP), 2021,