Biased ReLU neural networks

被引:32
|
作者
Liang, XingLong [1 ]
Xu, Jun [1 ]
机构
[1] Harbin Inst Technol, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Biased ReLU; Neural network; PWL network flexibility; ADAPTIVE HINGING HYPERPLANES;
D O I
10.1016/j.neucom.2020.09.050
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks (NN) with rectified linear units (ReLU) have been widely implemented since 2012. In this paper, we describe an activation function called the biased ReLU neuron (BReLU), which is similar to the ReLU. Based on this activation function, we propose the BReLU NN (BRNN). The structure of the BRNN is similar to that of the ReLU network. However, the difference between the two is that the BReLU introduces several biases for each input variable. This allows the BRNN to divide the input space into a greater number of linear regions and improve network flexibility. The BRNN parameters to be estimated are the weight matrices and the bias parameters of the BReLU neurons. The weights are obtained using the backpropagation method. Moreover, we propose a method to compute the bias parameters of the BReLU neurons. In this method, batch normalization is applied to the BRNN, and the variance and mean of the input variables are obtained. Based on these two parameters, the bias parameters are estimated. In addition, we investigate the flexibility of the BRNN. Specifically, we study the number of linear regions and provide the upper bound for the maximum number of linear regions. The results indicate that for the same input dimension, the BRNN divides the input space into a greater number of linear regions than the ReLU network. This explains to a certain extent why the BRNN has the superior approximation ability. Experiments are carried out using five datasets, and the results verify the effectiveness of the proposed method. (c) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页码:71 / 79
页数:9
相关论文
共 50 条
  • [1] Unboundedness of Linear Regions of Deep ReLU Neural Networks
    Ponomarchuk, Anton
    Koutschan, Christoph
    Moser, Bernhard
    DATABASE AND EXPERT SYSTEMS APPLICATIONS, DEXA 2022 WORKSHOPS, 2022, 1633 : 3 - 10
  • [2] ReLU, Sparseness, and the Encoding of Optic Flow in Neural Networks
    Layton, Oliver W.
    Peng, Siyuan
    Steinmetz, Scott T.
    SENSORS, 2024, 24 (23)
  • [3] Expressive Numbers of Two or More Hidden Layer ReLU Neural Networks
    Inoue, Kenta
    2019 SEVENTH INTERNATIONAL SYMPOSIUM ON COMPUTING AND NETWORKING WORKSHOPS (CANDARW 2019), 2019, : 129 - 135
  • [4] A Dynamic ReLU on Neural Network
    Si, Jiong
    Harris, Sarah L.
    Yfantis, Evangelos
    PROCEEDINGS OF THE 2018 IEEE 13TH DALLAS CIRCUITS AND SYSTEMS CONFERENCE (DCAS), 2018,
  • [5] Provable Identifiability of Two-Layer ReLU Neural Networks via LASSO Regularization
    Li G.
    Wang G.
    Ding J.
    IEEE Transactions on Information Theory, 2023, 69 (09) : 5921 - 5935
  • [6] Complexity of training ReLU neural network
    Boob, Digvijay
    Dey, Santanu S.
    Lan, Guanghui
    DISCRETE OPTIMIZATION, 2022, 44
  • [7] Convolutional Neural Network with Biologically Inspired ON/OFF ReLU
    Kim, Jonghong
    Kim, Seonggyu
    Lee, Minho
    NEURAL INFORMATION PROCESSING, ICONIP 2015, PT IV, 2015, 9492 : 316 - 323
  • [8] Quantile regression with ReLU Networks: Estimators and minimax rates
    Padilla, Oscar Hernan Madrid
    Tansey, Wesley
    Chen, Yanzhen
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [9] A multiparametric approach to accelerating ReLU neural network based model predictive control
    Kenefake, Dustin
    Kakodkar, Rahul
    Akundi, Sahithi S.
    Ali, Moustafa
    Pistikopoulos, Efstratios N.
    CONTROL ENGINEERING PRACTICE, 2024, 151
  • [10] Rule extraction: Using neural networks or for neural networks?
    Zhi-Hua Zhou
    Journal of Computer Science and Technology, 2004, 19 : 249 - 253